Bias by Design? How Data Practices Shape Fairness in AI Healthcare Systems
By: Anna Arias-Duart, Maria Eugenia Cardello, Atia Cortés
Potential Business Impact:
Makes medical AI fair for everyone.
Artificial intelligence (AI) holds great promise for transforming healthcare. However, despite significant advances, the integration of AI solutions into real-world clinical practice remains limited. A major barrier is the quality and fairness of training data, which is often compromised by biased data collection practices. This paper draws on insights from the AI4HealthyAging project, part of Spain's national R&D initiative, where our task was to detect biases during clinical data collection. We identify several types of bias across multiple use cases, including historical, representation, and measurement biases. These biases manifest in variables such as sex, gender, age, habitat, socioeconomic status, equipment, and labeling. We conclude with practical recommendations for improving the fairness and robustness of clinical problem design and data collection. We hope that our findings and experience contribute to guiding future projects in the development of fairer AI systems in healthcare.
Similar Papers
Datasheets for Healthcare AI: A Framework for Transparency and Bias Mitigation
Computers and Society
Makes AI in medicine fair and safe.
Beyond Internal Data: Constructing Complete Datasets for Fairness Testing
Machine Learning (CS)
Tests AI for fairness without private data.
AI Fairness Beyond Complete Demographics: Current Achievements and Future Directions
Computers and Society
Makes AI fair even with missing information.