What Makes Reasoning Invalid: Echo Reflection Mitigation for Large Language Models
By: Chen He , Xun Jiang , Lei Wang and more
Potential Business Impact:
Helps computers think deeper, not just repeat.
Large Language Models (LLMs) have demonstrated remarkable performance across a wide range of reasoning tasks. Recent methods have further improved LLM performance in complex mathematical reasoning. However, when extending these methods beyond the domain of mathematical reasoning to tasks involving complex domain-specific knowledge, we observe a consistent failure of LLMs to generate novel insights during the reflection stage. Instead of conducting genuine cognitive refinement, the model tends to mechanically reiterate earlier reasoning steps without introducing new information or perspectives, a phenomenon referred to as "Echo Reflection". We attribute this behavior to two key defects: (1) Uncontrollable information flow during response generation, which allows premature intermediate thoughts to propagate unchecked and distort final decisions; (2) Insufficient exploration of internal knowledge during reflection, leading to repeating earlier findings rather than generating new cognitive insights. Building on these findings, we proposed a novel reinforcement learning method termed Adaptive Entropy Policy Optimization (AEPO). Specifically, the AEPO framework consists of two major components: (1) Reflection-aware Information Filtration, which quantifies the cognitive information flow and prevents the final answer from being affected by earlier bad cognitive information; (2) Adaptive-Entropy Optimization, which dynamically balances exploration and exploitation across different reasoning stages, promoting both reflective diversity and answer correctness. Extensive experiments demonstrate that AEPO consistently achieves state-of-the-art performance over mainstream reinforcement learning baselines across diverse benchmarks.
Similar Papers
Illusions of reflection: open-ended task reveals systematic failures in Large Language Models' reflective reasoning
Artificial Intelligence
Computers don't learn from their own mistakes.
ReflCtrl: Controlling LLM Reflection via Representation Engineering
Artificial Intelligence
Control AI's thinking to save energy.
ReaLM: Reflection-Enhanced Autonomous Reasoning with Small Language Models
Computation and Language
Teaches small computers to think better on their own.