GuessingGame: Measuring the Informativeness of Open-Ended Questions in Large Language Models
By: Dylan Hutson , Daniel Vennemeyer , Aneesh Deshmukh and more
Potential Business Impact:
Teaches computers to ask smart questions to guess things.
We introduce GuessingGame, a protocol for evaluating large language models (LLMs) as strategic question-askers in open-ended, open-domain settings. A Guesser LLM identifies a hidden object by posing free-form questions to an Oracle without predefined choices or candidate lists. To measure question quality, we propose two information gain (IG) metrics: a Bayesian method that tracks belief updates over semantic concepts using LLM-scored relevance, and an entropy-based method that filters candidates via ConceptNet. Both metrics are model-agnostic and support post hoc analysis. Across 858 games with multiple models and prompting strategies, higher IG strongly predicts efficiency: a one-standard-deviation IG increase reduces expected game length by 43\%. Prompting constraints guided by IG, such as enforcing question diversity, enable weaker models to significantly improve performance. These results show that question-asking in LLMs is both measurable and improvable, and crucial for interactive reasoning.
Similar Papers
Measuring and Analyzing Intelligence via Contextual Uncertainty in Large Language Models using Information-Theoretic Metrics
Artificial Intelligence
Shows how AI thinks by tracking its guessing.
Cross-Entropy Games for Language Models: From Implicit Knowledge to General Capability Measures
Artificial Intelligence
Tests computers to see how smart they are.
Measuring Aleatoric and Epistemic Uncertainty in LLMs: Empirical Evaluation on ID and OOD QA Tasks
Computation and Language
Helps computers know when they are unsure.