CoVeR: Conformal Calibration for Versatile and Reliable Autoregressive Next-Token Prediction
By: Yuzhu Chen , Yingjie Wang , Shunyu Liu and more
Potential Business Impact:
Finds best answers with guaranteed success.
Autoregressive pre-trained models combined with decoding methods have achieved impressive performance on complex reasoning tasks. While mainstream decoding strategies such as beam search can generate plausible candidate sets, they often lack provable coverage guarantees, and struggle to effectively balance search efficiency with the need for versatile trajectories, particularly those involving long-tail sequences that are essential in certain real-world applications. To address these limitations, we propose \textsc{CoVeR}, a novel model-free decoding strategy wihtin the conformal prediction framework that simultaneously maintains a compact search space and ensures high coverage probability over desirable trajectories. Theoretically, we establish a PAC-style generalization bound, guaranteeing that \textsc{CoVeR} asymptotically achieves a coverage rate of at least $1 - \alpha$ for any target level $\alpha \in (0,1)$.
Similar Papers
Conformal Prediction Sets for Next-Token Prediction in Large Language Models: Balancing Coverage Guarantees with Set Efficiency
Computation and Language
Helps AI guess words more accurately and efficiently.
Conditional Coverage Diagnostics for Conformal Prediction
Machine Learning (Stat)
Helps computers know when their guesses are wrong.
Reliable Statistical Guarantees for Conformal Predictors with Small Datasets
Machine Learning (CS)
Makes AI predictions more trustworthy, even with little data.