Unveiling the Response of Large Vision-Language Models to Visually Absent Tokens
By: Sohee Kim , Soohyun Ryu , Joonhyung Park and more
Potential Business Impact:
Fixes AI that sees text in pictures.
Large Vision-Language Models (LVLMs) generate contextually relevant responses by jointly interpreting visual and textual inputs. However, our finding reveals they often mistakenly perceive text inputs lacking visual evidence as being part of the image, leading to erroneous responses. In light of this finding, we probe whether LVLMs possess an internal capability to determine if textual concepts are grounded in the image, and discover a specific subset of Feed-Forward Network (FFN) neurons, termed Visual Absence-aware (VA) neurons, that consistently signal the visual absence through a distinctive activation pattern. Leveraging these patterns, we develop a detection module that systematically classifies whether an input token is visually grounded. Guided by its prediction, we propose a method to refine the outputs by reinterpreting question prompts or replacing the detected absent tokens during generation. Extensive experiments show that our method effectively mitigates the models' tendency to falsely presume the visual presence of text input and its generality across various LVLMs.
Similar Papers
Unveiling the Response of Large Vision-Language Models to Visually Absent Tokens
CV and Pattern Recognition
Fixes AI that sees text in pictures.
Direct Visual Grounding by Directing Attention of Visual Tokens
CV and Pattern Recognition
Makes AI better at answering questions about pictures.
To Sink or Not to Sink: Visual Information Pathways in Large Vision-Language Models
CV and Pattern Recognition
Finds important image parts for AI understanding.