Inclusive, Differentially Private Federated Learning for Clinical Data
By: Santhosh Parampottupadam , Melih Coşğun , Sarthak Pati and more
Potential Business Impact:
Helps hospitals share patient data safely for better health.
Federated Learning (FL) offers a promising approach for training clinical AI models without centralizing sensitive patient data. However, its real-world adoption is hindered by challenges related to privacy, resource constraints, and compliance. Existing Differential Privacy (DP) approaches often apply uniform noise, which disproportionately degrades model performance, even among well-compliant institutions. In this work, we propose a novel compliance-aware FL framework that enhances DP by adaptively adjusting noise based on quantifiable client compliance scores. Additionally, we introduce a compliance scoring tool based on key healthcare and security standards to promote secure, inclusive, and equitable participation across diverse clinical settings. Extensive experiments on public datasets demonstrate that integrating under-resourced, less compliant clinics with highly regulated institutions yields accuracy improvements of up to 15% over traditional FL. This work advances FL by balancing privacy, compliance, and performance, making it a viable solution for real-world clinical workflows in global healthcare.
Similar Papers
An Interactive Framework for Implementing Privacy-Preserving Federated Learning: Experiments on Large Language Models
Machine Learning (CS)
Protects private data while training smart computer programs.
Differentially Private Federated Learning With Time-Adaptive Privacy Spending
Machine Learning (CS)
Learns more from private data, faster.
Optimal Strategies for Federated Learning Maintaining Client Privacy
Machine Learning (CS)
Makes private AI learning better without losing data.