Securing Genomic Data Against Inference Attacks in Federated Learning Environments
By: Chetan Pathade, Shubham Patil
Potential Business Impact:
Protects secret health codes from hackers.
Federated Learning (FL) offers a promising framework for collaboratively training machine learning models across decentralized genomic datasets without direct data sharing. While this approach preserves data locality, it remains susceptible to sophisticated inference attacks that can compromise individual privacy. In this study, we simulate a federated learning setup using synthetic genomic data and assess its vulnerability to three key attack vectors: Membership Inference Attack (MIA), Gradient-Based Membership Inference Attack, and Label Inference Attack (LIA). Our experiments reveal that Gradient-Based MIA achieves the highest effectiveness, with a precision of 0.79 and F1-score of 0.87, underscoring the risk posed by gradient exposure in federated updates. Additionally, we visualize comparative attack performance through radar plots and quantify model leakage across clients. The findings emphasize the inadequacy of na\"ive FL setups in safeguarding genomic privacy and motivate the development of more robust privacy-preserving mechanisms tailored to the unique sensitivity of genomic data.
Similar Papers
Exploring the Vulnerabilities of Federated Learning: A Deep Dive into Gradient Inversion Attacks
Cryptography and Security
Protects private info from sneaky computer learning.
An Efficient Gradient-Based Inference Attack for Federated Learning
Machine Learning (CS)
Finds private data hidden in shared learning updates.
Mitigating Membership Inference Vulnerability in Personalized Federated Learning
Machine Learning (CS)
Protects private data while improving AI learning.