Score: 1

Dropout Prompt Learning: Towards Robust and Adaptive Vision-Language Models

Published: December 8, 2025 | arXiv ID: 2512.07234v1

By: Biao Chen , Lin Zuo , Mengmeng Jing and more

Potential Business Impact:

Makes AI better at understanding new things.

Business Areas:
Natural Language Processing Artificial Intelligence, Data and Analytics, Software

Dropout is a widely used regularization technique which improves the generalization ability of a model by randomly dropping neurons. In light of this, we propose Dropout Prompt Learning, which aims for applying dropout to improve the robustness of the vision-language models. Different from the vanilla dropout, we apply dropout on the tokens of the textual and visual branches, where we evaluate the token significance considering both intra-modal context and inter-modal alignment, enabling flexible dropout probabilities for each token. Moreover, to maintain semantic alignment for general knowledge transfer while encouraging the diverse representations that dropout introduces, we further propose residual entropy regularization. Experiments on 15 benchmarks show our method's effectiveness in challenging scenarios like low-shot learning, long-tail classification, and out-of-distribution generalization. Notably, our method surpasses regularization-based methods including KgCoOp by 5.10% and PromptSRC by 2.13% in performance on base-to-novel generalization.

Country of Origin
🇨🇳 China

Repos / Data Links

Page Count
15 pages

Category
Computer Science:
CV and Pattern Recognition