Score: 3

Dynamic Thinking-Token Selection for Efficient Reasoning in Large Reasoning Models

Published: January 26, 2026 | arXiv ID: 2601.18383v1

By: Zhenyuan Guo , Tong Chen , Wenlong Meng and more

Potential Business Impact:

Makes AI think faster by skipping extra steps.

Business Areas:
Natural Language Processing Artificial Intelligence, Data and Analytics, Software

Large Reasoning Models (LRMs) excel at solving complex problems by explicitly generating a reasoning trace before deriving the final answer. However, these extended generations incur substantial memory footprint and computational overhead, bottlenecking LRMs' efficiency. This work uses attention maps to analyze the influence of reasoning traces and uncover an interesting phenomenon: only some decision-critical tokens in a reasoning trace steer the model toward the final answer, while the remaining tokens contribute negligibly. Building on this observation, we propose Dynamic Thinking-Token Selection (DynTS). This method identifies decision-critical tokens and retains only their associated Key-Value (KV) cache states during inference, evicting the remaining redundant entries to optimize efficiency.

Country of Origin
πŸ‡¨πŸ‡³ China


Page Count
18 pages

Category
Computer Science:
Artificial Intelligence