How Chain-of-Thought Works? Tracing Information Flow from Decoding, Projection, and Activation
By: Hao Yang, Qinghua Zhao, Lei Li
Potential Business Impact:
Makes AI think better by guiding its answers.
Chain-of-Thought (CoT) prompting significantly enhances model reasoning, yet its internal mechanisms remain poorly understood. We analyze CoT's operational principles by reversely tracing information flow across decoding, projection, and activation phases. Our quantitative analysis suggests that CoT may serve as a decoding space pruner, leveraging answer templates to guide output generation, with higher template adherence strongly correlating with improved performance. Furthermore, we surprisingly find that CoT modulates neuron engagement in a task-dependent manner: reducing neuron activation in open-domain tasks, yet increasing it in closed-domain scenarios. These findings offer a novel mechanistic interpretability framework and critical insights for enabling targeted CoT interventions to design more efficient and robust prompts. We released our code and data at https://anonymous.4open.science/r/cot-D247.
Similar Papers
Chain-of-Conceptual-Thought: Eliciting the Agent to Deeply Think within the Response
Computation and Language
Helps AI understand feelings and give better advice.
Understanding Chain-of-Thought Effectiveness in Code Generation: An Empirical and Information-Theoretic Analysis
Software Engineering
Helps computers write better code by thinking step-by-step.
The CoT Encyclopedia: Analyzing, Predicting, and Controlling how a Reasoning Model will Think
Computation and Language
Helps computers think better by understanding their thoughts.