Score: 2

SemVink: Advancing VLMs' Semantic Understanding of Optical Illusions via Visual Global Thinking

Published: June 3, 2025 | arXiv ID: 2506.02803v2

By: Sifan Li, Yujun Cai, Yiwei Wang

Potential Business Impact:

Makes computers see hidden things in pictures.

Business Areas:
Visual Search Internet Services

Vision-language models (VLMs) excel in semantic tasks but falter at a core human capability: detecting hidden content in optical illusions or AI-generated images through perceptual adjustments like zooming. We introduce HC-Bench, a benchmark of 112 images with hidden text, objects, and illusions, revealing that leading VLMs achieve near-zero accuracy (0-5.36%)-even with explicit prompting. Humans resolve such ambiguities instinctively, yet VLMs fail due to an overreliance on high-level semantics. Strikingly, we propose SemVink (Semantic Visual Thinking) by simply scaling images to low resolutions (32-128 pixels), which unlocks >99% accuracy by eliminating redundant visual noise. This exposes a critical architectural flaw: VLMs prioritize abstract reasoning over low-level visual operations crucial for real-world robustness. Our work urges a shift toward hybrid models integrating multi-scale processing, bridging the gap between computational vision and human cognition for applications in medical imaging, security, and beyond.

Country of Origin
πŸ‡¦πŸ‡Ί πŸ‡ΊπŸ‡Έ Australia, United States

Repos / Data Links

Page Count
10 pages

Category
Computer Science:
Computation and Language