Don't Miss the Forest for the Trees: In-Depth Confidence Estimation for LLMs via Reasoning over the Answer Space
By: Ante Wang, Weizhi Ma, Yang Liu
Potential Business Impact:
Helps AI know how sure it is about answers.
Knowing the reliability of a model's response is essential in application. With the strong generation capabilities of LLMs, research has focused on generating verbalized confidence. This is further enhanced by combining chain-of-thought reasoning, which provides logical and transparent estimation. However, how reasoning strategies affect the estimated confidence is still under-explored. In this work, we demonstrate that predicting a verbalized probability distribution can effectively encourage in-depth reasoning for confidence estimation. Intuitively, it requires an LLM to consider all candidates within the answer space instead of basing on a single guess, and to carefully assign confidence scores to meet the requirements of a distribution. This method shows an advantage across different models and various tasks, regardless of whether the answer space is known. Its advantage is maintained even after reinforcement learning, and further analysis shows its reasoning patterns are aligned with human expectations.
Similar Papers
Read Your Own Mind: Reasoning Helps Surface Self-Confidence Signals in LLMs
Computation and Language
Makes AI more honest about what it knows.
Open the Oyster: Empirical Evaluation and Improvement of Code Reasoning Confidence in LLMs
Software Engineering
Makes AI better at knowing when it's right.
Multiple Choice Questions: Reasoning Makes Large Language Models (LLMs) More Self-Confident Even When They Are Wrong
Computation and Language
Computers trust their answers more when they explain them.