State over Tokens: Characterizing the Role of Reasoning Tokens
By: Mosh Levy , Zohar Elyoseph , Shauli Ravfogel and more
Potential Business Impact:
Lets computers "think" better by showing their steps.
Large Language Models (LLMs) can generate reasoning tokens before their final answer to boost performance on complex tasks. While these sequences seem like human thought processes, empirical evidence reveals that they are not a faithful explanation of the model's actual reasoning process. To address this gap between appearance and function, we introduce the State over Tokens (SoT) conceptual framework. SoT reframes reasoning tokens not as a linguistic narrative, but as an externalized computational state -- the sole persistent information carrier across the model's stateless generation cycles. This explains how the tokens can drive correct reasoning without being a faithful explanation when read as text and surfaces previously overlooked research questions on these tokens. We argue that to truly understand the process that LLMs do, research must move beyond reading the reasoning tokens as text and focus on decoding them as state.
Similar Papers
LLM Reasoning for Machine Translation: Synthetic Data Generation over Thinking Tokens
Computation and Language
Makes computer translators better by showing them how.
Sketch-of-Thought: Efficient LLM Reasoning with Adaptive Cognitive-Inspired Sketching
Computation and Language
Makes smart computers think faster, using fewer words.
SoT: Structured-of-Thought Prompting Guides Multilingual Reasoning in Large Language Models
Computation and Language
Helps computers understand and reason in any language.