Score: 0

Context-Adaptive Multi-Prompt Embedding with Large Language Models for Vision-Language Alignment

Published: August 3, 2025 | arXiv ID: 2508.02762v2

By: Dahun Kim, Anelia Angelova

Potential Business Impact:

Makes computers understand pictures and words better.

We propose Context-Adaptive Multi-Prompt Embedding, a novel approach to enrich semantic representations in vision-language contrastive learning. Unlike standard CLIP-style models that rely on a single text embedding, our method introduces multiple structured prompts, each containing a distinct adaptive token that captures diverse semantic aspects of the input text. We leverage a pretrained LLM as the text encoder within the CLIP framework, processing all prompts jointly in a single forward pass. The resulting prompt embeddings are combined into a unified text representation, enabling semantically richer alignment with visual features. To further promote semantic diversity and representation quality, we incorporate a diversity regularization loss and a negation-aware loss, encouraging specialization across prompts and improving contrastive discrimination. Our method achieves consistent improvements on both image-text and video-text retrieval benchmarks.

Page Count
13 pages

Category
Computer Science:
Machine Learning (CS)