How Efficient Are Diffusion Language Models? A Critical Examination of Efficiency Evaluation Practices
By: Han Peng , Peiyu Liu , Zican Dong and more
Potential Business Impact:
Makes AI models learn and create faster.
Diffusion language models (DLMs) have emerged as a promising alternative to the long-dominant autoregressive (AR) paradigm, offering a parallelable decoding process that could yield greater efficiency. Yet, in practice, current open-source DLMs often underperform their AR counterparts in speed, limiting their real-world utility. This work presents a systematic study of DLM efficiency, identifying key issues in prior evaluation methods. Through empirical benchmarking and a roofline-based theoretical analysis, we demonstrate that AR models generally achieve higher throughput, while DLMs consistently lag. We also investigate acceleration strategies, finding that techniques like dual cache and parallel decoding mainly offer gains at small batch sizes, with their benefits diminishing upon scaling. Our findings underscore the necessity of robust evaluation methods and improved acceleration strategies to advance research on DLMs.
Similar Papers
A Survey on Diffusion Language Models
Computation and Language
Makes computers write faster and understand better.
Beyond Next-Token Prediction: A Performance Characterization of Diffusion versus Autoregressive Language Models
Machine Learning (CS)
Makes computers write faster and understand longer stories.
Diffusion Language Models are Super Data Learners
Machine Learning (CS)
Makes AI better at writing code with less data.