Score: 0

Reference Games as a Testbed for the Alignment of Model Uncertainty and Clarification Requests

Published: January 12, 2026 | arXiv ID: 2601.07820v1

By: Manar Ali , Judith Sieker , Sina Zarrieß and more

Potential Business Impact:

Models ask for help when they don't understand.

Business Areas:
Semantic Search Internet Services

In human conversation, both interlocutors play an active role in maintaining mutual understanding. When addressees are uncertain about what speakers mean, for example, they can request clarification. It is an open question for language models whether they can assume a similar addressee role, recognizing and expressing their own uncertainty through clarification. We argue that reference games are a good testbed to approach this question as they are controlled, self-contained, and make clarification needs explicit and measurable. To test this, we evaluate three vision-language models comparing a baseline reference resolution task to an experiment where the models are instructed to request clarification when uncertain. The results suggest that even in such simple tasks, models often struggle to recognize internal uncertainty and translate it into adequate clarification behavior. This demonstrates the value of reference games as testbeds for interaction qualities of (vision and) language models.

Page Count
9 pages

Category
Computer Science:
Computation and Language