Unveiling Intrinsic Text Bias in Multimodal Large Language Models through Attention Key-Space Analysis
By: Xinhan Zheng , Huyu Wu , Xueting Wang and more
Potential Business Impact:
Fixes AI's bias towards words over pictures.
Multimodal large language models (MLLMs) exhibit a pronounced preference for textual inputs when processing vision-language data, limiting their ability to reason effectively from visual evidence. Unlike prior studies that attribute this text bias to external factors such as data imbalance or instruction tuning, we propose that the bias originates from the model's internal architecture. Specifically, we hypothesize that visual key vectors (Visual Keys) are out-of-distribution (OOD) relative to the text key space learned during language-only pretraining. Consequently, these visual keys receive systematically lower similarity scores during attention computation, leading to their under-utilization in the context representation. To validate this hypothesis, we extract key vectors from LLaVA and Qwen2.5-VL and analyze their distributional structures using qualitative (t-SNE) and quantitative (Jensen-Shannon divergence) methods. The results provide direct evidence that visual and textual keys occupy markedly distinct subspaces within the attention space. The inter-modal divergence is statistically significant, exceeding intra-modal variation by several orders of magnitude. These findings reveal that text bias arises from an intrinsic misalignment within the attention key space rather than solely from external data factors.
Similar Papers
See What You Are Told: Visual Attention Sink in Large Multimodal Models
CV and Pattern Recognition
Makes AI better at looking at pictures.
When Language Overrules: Revealing Text Dominance in Multimodal Large Language Models
Computation and Language
Makes AI use all senses, not just reading.
Exploring Implicit Visual Misunderstandings in Multimodal Large Language Models through Attention Analysis
CV and Pattern Recognition
Checks if AI truly sees pictures, not just guesses.