VLMs have Tunnel Vision: Evaluating Nonlocal Visual Reasoning in Leading VLMs
By: Shmuel Berman, Jia Deng
Potential Business Impact:
Computers can't connect image parts like humans.
Visual Language Models (VLMs) excel at complex visual tasks such as VQA and chart understanding, yet recent work suggests they struggle with simple perceptual tests. We present an evaluation that tests vision-language models' capacity for nonlocal visual reasoning -- reasoning that requires chaining evidence collected from multiple, possibly distant, regions of an image. We isolate three distinct forms of non-local vision: comparative perception, which demands holding two images in working memory and comparing them; saccadic search, which requires making discrete, evidence-driven jumps to locate successive targets; and smooth visual search, which involves searching smoothly along a continuous contour. Flagship models (e.g., Gemini 2.5 Pro, Claude Vision 3.7, GPT-o4-mini), even those that perform well on prior primitive-vision benchmarks, fail these tests and barely exceed random accuracy on two variants of our tasks that are trivial for humans. Our structured evaluation suite allows us to test if VLMs can perform similar visual algorithms to humans. Our findings show that despite gains in raw visual acuity, current models lack core visual reasoning capabilities.
Similar Papers
Caption This, Reason That: VLMs Caught in the Middle
CV and Pattern Recognition
Helps computers understand pictures better by thinking.
Visual Language Models show widespread visual deficits on neuropsychological tests
CV and Pattern Recognition
Computers see things like humans, but miss basic details.
Vision language models are unreliable at trivial spatial cognition
CV and Pattern Recognition
Computers struggle to tell what's left or right.