Robust Clustered Federated Learning for Heterogeneous High-dimensional Data
By: Changxin Yang, Zhongyi Zhu, Heng Lian
Potential Business Impact:
Helps computers learn from private data better.
Federated learning has attracted significant attention as a privacy-preserving framework for training personalised models on multi-source heterogeneous data. However, most existing approaches are unable to handle scenarios where subgroup structures coexist alongside within-group heterogeneity. In this paper, we propose a federated learning algorithm that addresses general heterogeneity through adaptive clustering. Specifically, our method partitions tasks into subgroups to address substantial between-group differences while enabling efficient information sharing among similar tasks within each group. Furthermore, we integrate the Huber loss and Iterative Hard Thresholding (IHT) to tackle the challenges of high dimensionality and heavy-tailed distributions. Theoretically, we establish convergence guarantees, derive non-asymptotic error bounds, and provide recovery guarantees for the latent cluster structure. Extensive simulation studies and real-data applications further demonstrate the effectiveness and adaptability of our approach.
Similar Papers
Federated Online Learning for Heterogeneous Multisource Streaming Data
Machine Learning (Stat)
Learns from many computers without sharing private data.
FedHFT: Efficient Federated Finetuning with Heterogeneous Edge Clients
Machine Learning (CS)
Helps AI learn from private data securely.
Federated Learning Framework for Scalable AI in Heterogeneous HPC and Cloud Environments
Distributed, Parallel, and Cluster Computing
Trains AI on many computers without sharing private data.