Diffusion Language Model Inference with Monte Carlo Tree Search
By: Zheng Huang , Kiran Ramnath , Yueyan Chen and more
Potential Business Impact:
Makes AI write better by finding best word choices.
Diffusion language models (DLMs) have recently emerged as a compelling alternative to autoregressive generation, offering parallel generation and improved global coherence. During inference, DLMs generate text by iteratively denoising masked sequences in parallel; however, determining which positions to unmask and which tokens to commit forms a large combinatorial search problem. Existing inference methods approximate this search using heuristics, which often yield suboptimal decoding paths; other approaches instead rely on additional training to guide token selection. To introduce a principled search mechanism for DLMs inference, we introduce MEDAL, a framework that integrates Monte Carlo Tree SEarch initialization for Diffusion LAnguage Model inference. We employ Monte Carlo Tree Search at the initialization stage to explore promising unmasking trajectories, providing a robust starting point for subsequent refinement. This integration is enabled by restricting the search space to high-confidence actions and prioritizing token choices that improve model confidence over remaining masked positions. Across multiple benchmarks, MEDAL achieves up to 22.0% improvement over existing inference strategies, establishing a new paradigm for search-based inference in diffusion language models.
Similar Papers
A Survey on Diffusion Language Models
Computation and Language
Makes computers write faster and understand better.
Diffusion Language Models Know the Answer Before Decoding
Computation and Language
Makes AI answer questions much faster.
No Compute Left Behind: Rethinking Reasoning and Sampling with Masked Diffusion Models
Machine Learning (CS)
Helps computers solve math problems better.