Entropy-Aligned Decoding of LMs for Better Writing and Reasoning
By: Kareem Ahmed, Sameer Singh
Potential Business Impact:
Makes AI write better stories and answers.
Language models (LMs) are trained on billions of tokens in an attempt to recover the true language distribution. Still, vanilla random sampling from LMs yields low quality generations. Decoding algorithms attempt to restrict the LM distribution to a set of high-probability continuations, but rely on greedy heuristics that introduce myopic distortions, yielding sentences that are homogeneous, repetitive and incoherent. In this paper, we introduce EPIC, a hyperparameter-free decoding approach that incorporates the entropy of future trajectories into LM decoding. EPIC explicitly regulates the amount of uncertainty expressed at every step of generation, aligning the sampling distribution's entropy to the aleatoric (data) uncertainty. Through Entropy-Aware Lazy Gumbel-Max sampling, EPIC manages to be exact, while also being efficient, requiring only a sublinear number of entropy evaluations per step. Unlike current baselines, EPIC yields sampling distributions that are empirically well-aligned with the entropy of the underlying data distribution. Across creative writing and summarization tasks, EPIC consistently improves LM-as-judge preference win-rates over widely used decoding strategies. These preference gains are complemented by automatic metrics, showing that EPIC produces more diverse generations and more faithful summaries. We also evaluate EPIC on mathematical reasoning, where it outperforms all baselines.
Similar Papers
GEM: Generative Entropy-Guided Preference Modeling for Few-shot Alignment of LLMs
Artificial Intelligence
Teaches AI to learn from expert opinions.
Reasoning Planning for Language Models
Machine Learning (CS)
Helps computers pick the best way to solve math problems.
From Bits to Rounds: Parallel Decoding with Exploration for Diffusion Language Models
Machine Learning (CS)
Makes AI write faster by finding better words.