Generalization in Federated Learning: A Conditional Mutual Information Framework
By: Ziqiao Wang, Cheng Long, Yongyi Mao
Potential Business Impact:
Improves how computers learn from many sources privately.
Federated learning (FL) is a widely adopted privacy-preserving distributed learning framework, yet its generalization performance remains less explored compared to centralized learning. In FL, the generalization error consists of two components: the out-of-sample gap, which measures the gap between the empirical and true risk for participating clients, and the participation gap, which quantifies the risk difference between participating and non-participating clients. In this work, we apply an information-theoretic analysis via the conditional mutual information (CMI) framework to study FL's two-level generalization. Beyond the traditional supersample-based CMI framework, we introduce a superclient construction to accommodate the two-level generalization setting in FL. We derive multiple CMI-based bounds, including hypothesis-based CMI bounds, illustrating how privacy constraints in FL can imply generalization guarantees. Furthermore, we propose fast-rate evaluated CMI bounds that recover the best-known convergence rate for two-level FL generalization in the small empirical risk regime. For specific FL model aggregation strategies and structured loss functions, we refine our bounds to achieve improved convergence rates with respect to the number of participating clients. Empirical evaluations confirm that our evaluated CMI bounds are non-vacuous and accurately capture the generalization behavior of FL algorithms.
Similar Papers
Benchmarking Mutual Information-based Loss Functions in Federated Learning
Machine Learning (CS)
Makes AI fairer for everyone using less data.
A Generalized Meta Federated Learning Framework with Theoretical Convergence Guarantees
Machine Learning (CS)
Helps AI learn better from many separate computers.
Local Performance vs. Out-of-Distribution Generalization: An Empirical Analysis of Personalized Federated Learning in Heterogeneous Data Environments
Machine Learning (CS)
Helps AI learn better from different data.