Calibration-Aware Prompt Learning for Medical Vision-Language Models
By: Abhishek Basu , Fahad Shamshad , Ashshak Sharifdeen and more
Potential Business Impact:
Makes AI doctors more sure about their guesses.
Medical Vision-Language Models (Med-VLMs) have demonstrated remarkable performance across diverse medical imaging tasks by leveraging large-scale image-text pretraining. However, their confidence calibration is largely unexplored, and so remains a significant challenge. As such, miscalibrated predictions can lead to overconfident errors, undermining clinical trust and decision-making reliability. To address this, we introduce CalibPrompt, the first framework to calibrate Med-VLMs during prompt tuning. CalibPrompt optimizes a small set of learnable prompts with carefully designed calibration objectives under scarce labeled data regime. First, we study a regularizer that attempts to align the smoothed accuracy with the predicted model confidences. Second, we introduce an angular separation loss to maximize textual feature proximity toward improving the reliability in confidence estimates of multimodal Med-VLMs. Extensive experiments on four publicly available Med-VLMs and five diverse medical imaging datasets reveal that CalibPrompt consistently improves calibration without drastically affecting clean accuracy. Our code is available at https://github.com/iabh1shekbasu/CalibPrompt.
Similar Papers
Medical Knowledge Intervention Prompt Tuning for Medical Image Classification
CV and Pattern Recognition
Helps AI understand medical images better.
DualPrompt-MedCap: A Dual-Prompt Enhanced Approach for Medical Image Captioning
CV and Pattern Recognition
Helps doctors understand medical pictures better.
VaMP: Variational Multi-Modal Prompt Learning for Vision-Language Models
CV and Pattern Recognition
Teaches computers to understand pictures and words better.