Single-Round Clustered Federated Learning via Data Collaboration Analysis for Non-IID Data
By: Sota Sugawara , Yuji Kawamata , Akihiro Toyoda and more
Federated Learning (FL) enables distributed learning across multiple clients without sharing raw data. When statistical heterogeneity across clients is severe, Clustered Federated Learning (CFL) can improve performance by grouping similar clients and training cluster-wise models. However, most CFL approaches rely on multiple communication rounds for cluster estimation and model updates, which limits their practicality under tight constraints on communication rounds. We propose Data Collaboration-based Clustered Federated Learning (DC-CFL), a single-round framework that completes both client clustering and cluster-wise learning, using only the information shared in DC analysis. DC-CFL quantifies inter-client similarity via total variation distance between label distributions, estimates clusters using hierarchical clustering, and performs cluster-wise learning via DC analysis. Experiments on multiple open datasets under representative non-IID conditions show that DC-CFL achieves accuracy comparable to multi-round baselines while requiring only one communication round. These results indicate that DC-CFL is a practical alternative for collaborative AI model development when multiple communication rounds are impractical.
Similar Papers
A new type of federated clustering: A non-model-sharing approach
Machine Learning (CS)
Lets groups learn from private data together.
A Survey on Cluster-based Federated Learning
Machine Learning (Stat)
Groups computers to learn better from different data.
LCFed: An Efficient Clustered Federated Learning Framework for Heterogeneous Data
Machine Learning (CS)
Makes AI learn better from different groups of data.