Score: 3

Context informs pragmatic interpretation in vision-language models

Published: November 5, 2025 | arXiv ID: 2511.03908v1

By: Alvin Wei Ming Tan , Ben Prystawski , Veronica Boyce and more

BigTech Affiliations: Stanford University

Potential Business Impact:

Computers learn to understand context like people.

Business Areas:
Semantic Search Internet Services

Iterated reference games - in which players repeatedly pick out novel referents using language - present a test case for agents' ability to perform context-sensitive pragmatic reasoning in multi-turn linguistic environments. We tested humans and vision-language models on trials from iterated reference games, varying the given context in terms of amount, order, and relevance. Without relevant context, models were above chance but substantially worse than humans. However, with relevant context, model performance increased dramatically over trials. Few-shot reference games with abstract referents remain a difficult task for machine learning models.

Country of Origin
🇺🇸 United States

Repos / Data Links

Page Count
13 pages

Category
Computer Science:
Computation and Language