Score: 0

Referential ambiguity and clarification requests: comparing human and LLM behaviour

Published: July 14, 2025 | arXiv ID: 2507.10445v1

By: Chris Madge, Matthew Purver, Massimo Poesio

Potential Business Impact:

Helps computers ask better questions when confused.

Business Areas:
Semantic Search Internet Services

In this work we examine LLMs' ability to ask clarification questions in task-oriented dialogues that follow the asynchronous instruction-giver/instruction-follower format. We present a new corpus that combines two existing annotations of the Minecraft Dialogue Corpus -- one for reference and ambiguity in reference, and one for SDRT including clarifications -- into a single common format providing the necessary information to experiment with clarifications and their relation to ambiguity. With this corpus we compare LLM actions with original human-generated clarification questions, examining how both humans and LLMs act in the case of ambiguity. We find that there is only a weak link between ambiguity and humans producing clarification questions in these dialogues, and low correlation between humans and LLMs. Humans hardly ever produce clarification questions for referential ambiguity, but often do so for task-based uncertainty. Conversely, LLMs produce more clarification questions for referential ambiguity, but less so for task uncertainty. We question if LLMs' ability to ask clarification questions is predicated on their recent ability to simulate reasoning, and test this with different reasoning approaches, finding that reasoning does appear to increase question frequency and relevancy.

Country of Origin
πŸ‡¬πŸ‡§ United Kingdom

Page Count
11 pages

Category
Computer Science:
Computation and Language