Parallel Thinking, Sequential Answering: Bridging NAR and AR for Efficient Reasoning
By: Qihang Ai, Haiyun Jiang
Potential Business Impact:
Makes computers solve hard problems much faster.
We study reasoning tasks through a framework that integrates auto-regressive (AR) and non-autoregressive (NAR) language models. AR models, which generate text sequentially, excel at producing coherent outputs but often suffer from slow inference, particularly in reasoning-intensive domains such as mathematics and code, where lengthy chains of thought are required. In contrast, NAR models, such as discrete diffusion models, allow parallel generation and offer substantial speedups, though typically at the cost of reduced output quality. To address these limitations, we introduce a new paradigm in which an NAR model efficiently produces intermediate reasoning traces, which subsequently guide an AR model to deliver precise final answers. Experiments demonstrate that our approach yields significant 26% improvements over strong baselines while substantially reducing inference cost.
Similar Papers
SDAR: A Synergistic Diffusion-AutoRegression Paradigm for Scalable Sequence Generation
Machine Learning (CS)
Makes AI think faster and better.
TiDAR: Think in Diffusion, Talk in Autoregression
Computation and Language
Makes computers write better and faster.
Pseudo-Autoregressive Neural Codec Language Models for Efficient Zero-Shot Text-to-Speech Synthesis
Audio and Speech Processing
Makes computers talk like people faster.