Score: 1

VLMs have Tunnel Vision: Evaluating Nonlocal Visual Reasoning in Leading VLMs

Published: July 4, 2025 | arXiv ID: 2507.13361v1

By: Shmuel Berman, Jia Deng

BigTech Affiliations: Princeton University

Potential Business Impact:

Computers can't connect image parts like humans.

Business Areas:
Visual Search Internet Services

Visual Language Models (VLMs) excel at complex visual tasks such as VQA and chart understanding, yet recent work suggests they struggle with simple perceptual tests. We present an evaluation that tests vision-language models' capacity for nonlocal visual reasoning -- reasoning that requires chaining evidence collected from multiple, possibly distant, regions of an image. We isolate three distinct forms of non-local vision: comparative perception, which demands holding two images in working memory and comparing them; saccadic search, which requires making discrete, evidence-driven jumps to locate successive targets; and smooth visual search, which involves searching smoothly along a continuous contour. Flagship models (e.g., Gemini 2.5 Pro, Claude Vision 3.7, GPT-o4-mini), even those that perform well on prior primitive-vision benchmarks, fail these tests and barely exceed random accuracy on two variants of our tasks that are trivial for humans. Our structured evaluation suite allows us to test if VLMs can perform similar visual algorithms to humans. Our findings show that despite gains in raw visual acuity, current models lack core visual reasoning capabilities.

Country of Origin
🇺🇸 United States

Page Count
20 pages

Category
Computer Science:
CV and Pattern Recognition