Fill the Gap: Quantifying and Reducing the Modality Gap in Image-Text Representation Learning
By: François Role, Sébastien Meyer, Victor Amblard
Potential Business Impact:
Fixes how computers understand pictures and words together.
Vision-language models (VLMs) allow to embed texts and images in a shared representation space. However, it has been shown that these models are subject to a modality gap phenomenon meaning there exists a clear separation between the embeddings from one modality and another in the embedding space. While this misalignment is detrimental for downstream tasks such as multimodal retrieval, multimodal clustering or zero-shot classification, etc. no generic and practical methods have so far been proposed to assess it precisely and even reduce it. We therefore propose novel measures and effective techniques (spectral- and optimal transport-based methods) to achieve this goal. Extensive experiments conducted on several image-text datasets and models demonstrate their effectiveness and beneficial effects on downstream tasks. Our code is available at the URL provided in the paper's abstract.
Similar Papers
Exploring Textual Semantics Diversity for Image Transmission in Semantic Communication Systems using Visual Language Model
CV and Pattern Recognition
Sends pictures better by describing them with words.
Bridging the Modality Gap by Similarity Standardization with Pseudo-Positive Samples
Computation and Language
Makes searching text and pictures together work better.
Understanding the Modality Gap: An Empirical Study on the Speech-Text Alignment Mechanism of Large Speech Language Models
Computation and Language
Helps computers understand spoken words better.