Decoding Uncertainty: The Impact of Decoding Strategies for Uncertainty Estimation in Large Language Models
By: Wataru Hashimoto, Hidetaka Kamigaito, Taro Watanabe
Potential Business Impact:
Makes AI guess better and know when it's unsure.
Decoding strategies manipulate the probability distribution underlying the output of a language model and can therefore affect both generation quality and its uncertainty. In this study, we investigate the impact of decoding strategies on uncertainty estimation in Large Language Models (LLMs). Our experiments show that Contrastive Search, which mitigates repetition, yields better uncertainty estimates on average across a range of preference-aligned LLMs. In contrast, the benefits of these strategies sometimes diverge when the model is only post-trained with supervised fine-tuning, i.e. without explicit alignment.
Similar Papers
A Comparative Study of Decoding Strategies in Medical Text Generation
Computation and Language
Improves AI's medical answers by choosing the best words.
Towards Better Code Generation: Adaptive Decoding with Uncertainty Guidance
Software Engineering
Makes computers write better code by fixing mistakes.
A Survey of Uncertainty Estimation Methods on Large Language Models
Computation and Language
Helps AI tell when it's making things up.