Dropout Prompt Learning: Towards Robust and Adaptive Vision-Language Models
By: Biao Chen , Lin Zuo , Mengmeng Jing and more
Potential Business Impact:
Makes AI better at understanding new things.
Dropout is a widely used regularization technique which improves the generalization ability of a model by randomly dropping neurons. In light of this, we propose Dropout Prompt Learning, which aims for applying dropout to improve the robustness of the vision-language models. Different from the vanilla dropout, we apply dropout on the tokens of the textual and visual branches, where we evaluate the token significance considering both intra-modal context and inter-modal alignment, enabling flexible dropout probabilities for each token. Moreover, to maintain semantic alignment for general knowledge transfer while encouraging the diverse representations that dropout introduces, we further propose residual entropy regularization. Experiments on 15 benchmarks show our method's effectiveness in challenging scenarios like low-shot learning, long-tail classification, and out-of-distribution generalization. Notably, our method surpasses regularization-based methods including KgCoOp by 5.10% and PromptSRC by 2.13% in performance on base-to-novel generalization.
Similar Papers
Decoupling Augmentation Bias in Prompt Learning for Vision-Language Models
CV and Pattern Recognition
Teaches computers to recognize new things better.
Causality-guided Prompt Learning for Vision-language Models via Visual Granulation
CV and Pattern Recognition
Teaches computers to tell tiny differences between things.
Unleashing the Power of Vision-Language Models for Long-Tailed Multi-Label Visual Recognition
CV and Pattern Recognition
Helps computers see many things in pictures.