A Statistical Physics of Language Model Reasoning
By: Jack David Carson, Amir Reisizadeh
Potential Business Impact:
Explains how AI thinks, predicts mistakes.
Transformer LMs show emergent reasoning that resists mechanistic understanding. We offer a statistical physics framework for continuous-time chain-of-thought reasoning dynamics. We model sentence-level hidden state trajectories as a stochastic dynamical system on a lower-dimensional manifold. This drift-diffusion system uses latent regime switching to capture diverse reasoning phases, including misaligned states or failures. Empirical trajectories (8 models, 7 benchmarks) show a rank-40 projection (balancing variance capture and feasibility) explains ~50% variance. We find four latent reasoning regimes. An SLDS model is formulated and validated to capture these features. The framework enables low-cost reasoning simulation, offering tools to study and predict critical transitions like misaligned states or other LM failures.
Similar Papers
SwiReasoning: Switch-Thinking in Latent and Explicit for Pareto-Superior Reasoning LLMs
Computation and Language
Helps computers solve problems faster and smarter.
Explainable Chain-of-Thought Reasoning: An Empirical Analysis on State-Aware Reasoning Dynamics
Computation and Language
Shows how computers think step-by-step.
Have Large Language Models Learned to Reason? A Characterization via 3-SAT Phase Transition
Artificial Intelligence
Helps computers truly think, not just guess.