Optimizing Decoding Paths in Masked Diffusion Models by Quantifying Uncertainty
By: Ziyu Chen , Xinbei Jiang , Peng Sun and more
Potential Business Impact:
Finds better answers by managing guessing.
Masked Diffusion Models (MDMs) offer flexible, non-autoregressive generation, but this freedom introduces a challenge: final output quality is highly sensitive to the decoding order. We are the first to formalize this issue, attributing the variability in output quality to the cumulative predictive uncertainty along a generative path. To quantify this uncertainty, we introduce Denoising Entropy, a computable metric that serves as an internal signal for evaluating generative process. Leveraging this metric, we propose two algorithms designed to optimize the decoding path: a post-hoc selection method and a real-time guidance strategy. Experiments demonstrate that our entropy-guided methods significantly improve generation quality, consistently boosting accuracy on challenging reasoning, planning, and code benchmarks. Our work establishes Denoising Entropy as a principled tool for understanding and controlling generation, effectively turning the uncertainty in MDMs from a liability into a key advantage for discovering high-quality solutions.
Similar Papers
Masked Diffusion Models are Secretly Learned-Order Autoregressive Models
Machine Learning (CS)
Teaches computers to create ordered text better.
Improving Discrete Diffusion Unmasking Policies Beyond Explicit Reference Policies
Machine Learning (CS)
Teaches computers to write better sentences.
Error Bounds and Optimal Schedules for Masked Diffusions with Factorized Approximations
Machine Learning (Stat)
Makes AI create better text faster.