LLMs Have a Heart of Stone: Demystifying the Soft Thinking Ability of Large Reasoning Models
By: Junhong Wu , Jinliang Lu , Zixuan Ren and more
Potential Business Impact:
Makes AI think more creatively and solve problems better.
Human cognition naturally engages with abstract and fluid concepts, whereas existing reasoning models often rely on generating discrete tokens, potentially constraining their expressive capabilities. Recent advancements aim to address this limitation by enabling large language models (LLMs) to generate soft, abstract tokens, thus facilitating reasoning within a continuous concept space. This paper explores the `Soft Thinking' capabilities of various LLMs by examining the models' internal behavior using a suite of probing techniques. Contrary to the common belief that Soft Thinking enables the simultaneous exploration of diverse reasoning paths, our findings reveal that LLMs predominantly rely on the most influential component of the soft inputs during subsequent decoding steps. This reliance hinders the exploration of different reasoning paths and reduces vanilla Soft Thinking to a form of greedy decoding, obscuring the advantage of transmitting more information through Soft Tokens. To tackle this issue, we explore sampling strategies to introduce \emph{randomness}, employing methods such as Dirichlet resampling and the Gumbel-Softmax trick. Our experiments demonstrate that incorporating randomness can alleviate the limitations of vanilla approaches and unleash the potential of Soft Thinking. Notably, the Gumbel-Softmax trick provides adequate randomness with controlled smoothness, resulting in superior performance across eight reasoning benchmarks.
Similar Papers
LLMs are Single-threaded Reasoners: Demystifying the Working Mechanism of Soft Thinking
Computation and Language
Makes AI think more creatively and solve problems better.
Improving Latent Reasoning in LLMs via Soft Concept Mixing
Computation and Language
Teaches computers to think with fuzzy ideas.
SwiReasoning: Switch-Thinking in Latent and Explicit for Pareto-Superior Reasoning LLMs
Computation and Language
Helps computers solve problems faster and smarter.