Fairness in Multi-modal Medical Diagnosis with Demonstration Selection
By: Dawei Li , Zijian Gu , Peng Wang and more
Potential Business Impact:
Makes AI see medical images fairly for everyone.
Multimodal large language models (MLLMs) have shown strong potential for medical image reasoning, yet fairness across demographic groups remains a major concern. Existing debiasing methods often rely on large labeled datasets or fine-tuning, which are impractical for foundation-scale models. We explore In-Context Learning (ICL) as a lightweight, tuning-free alternative for improving fairness. Through systematic analysis, we find that conventional demonstration selection (DS) strategies fail to ensure fairness due to demographic imbalance in selected exemplars. To address this, we propose Fairness-Aware Demonstration Selection (FADS), which builds demographically balanced and semantically relevant demonstrations via clustering-based sampling. Experiments on multiple medical imaging benchmarks show that FADS consistently reduces gender-, race-, and ethnicity-related disparities while maintaining strong accuracy, offering an efficient and scalable path toward fair medical image reasoning. These results highlight the potential of fairness-aware in-context learning as a scalable and data-efficient solution for equitable medical image reasoning.
Similar Papers
Fairness in Multi-modal Medical Diagnosis with Demonstration Selection
CV and Pattern Recognition
Makes AI see medical images fairly for everyone.
MultiFair: Multimodal Balanced Fairness-Aware Medical Classification with Dual-Level Gradient Modulation
Machine Learning (CS)
Makes medical AI fairer and more accurate.
BiasICL: In-Context Learning and Demographic Biases of Vision Language Models
CV and Pattern Recognition
Helps AI doctors see health differences in people.