LLM-empowered Dynamic Prompt Routing for Vision-Language Models Tuning under Long-Tailed Distributions
By: Yongju Jia , Jiarui Ma , Xiangxian Li and more
Potential Business Impact:
Fixes computer vision bias in messy data.
Pre-trained vision-language models (VLMs), such as CLIP, have demonstrated impressive capability in visual tasks, but their fine-tuning often suffers from bias in class-imbalanced scene. Recent works have introduced large language models (LLMs) to enhance VLM fine-tuning with supplementing semantic information. However, they often overlook inherent class imbalance in VLMs' pre-training, which may lead to bias accumulation in downstream tasks. To address this problem, this paper proposes a Multi-dimensional Dynamic Prompt Routing (MDPR) framework. MDPR constructs a comprehensive knowledge base for classes, spanning five visual-semantic dimensions. During fine-tuning, the dynamic routing mechanism aligns global visual classes, retrieves optimal prompts, and balances fine-grained semantics, yielding stable predictions through logits fusion. Extensive experiments on long-tailed benchmarks, including CIFAR-LT, ImageNet-LT, and Places-LT, demonstrate that MDPR achieves comparable results with current SOTA methods. Ablation studies further confirm the effectiveness of our semantic library for tail classes, and show that our dynamic routing incurs minimal computational overhead, making MDPR a flexible and efficient enhancement for VLM fine-tuning under data imbalance.
Similar Papers
Generalizing Vision-Language Models with Dedicated Prompt Guidance
CV and Pattern Recognition
Helps AI understand new things better.
VaMP: Variational Multi-Modal Prompt Learning for Vision-Language Models
CV and Pattern Recognition
Teaches computers to understand pictures and words better.
Unleashing the Power of Vision-Language Models for Long-Tailed Multi-Label Visual Recognition
CV and Pattern Recognition
Helps computers see many things in pictures.