Asynchronous Federated Learning with non-convex client objective functions and heterogeneous dataset
By: Ali Forootani, Raffaele Iervolino
Potential Business Impact:
Trains AI faster without sharing private data.
Federated Learning (FL) enables collaborative model training across decentralized devices while preserving data privacy. However, traditional FL suffers from communication overhead, system heterogeneity, and straggler effects. Asynchronous Federated Learning (AFL) addresses these by allowing clients to update independently, improving scalability and reducing synchronization delays. This paper extends AFL to handle non-convex objective functions and heterogeneous datasets, common in modern deep learning. We present a rigorous convergence analysis, deriving bounds on the expected gradient norm and studying the effects of staleness, variance, and heterogeneity. To mitigate stale updates, we introduce a staleness aware aggregation that prioritizes fresher updates and a dynamic learning rate schedule that adapts to client staleness and heterogeneity, improving stability and convergence. Our framework accommodates variations in computational power, data distribution, and communication delays, making it practical for real world applications. We also analyze the impact of client selection strategies-sampling with or without replacement-on variance and convergence. Implemented in PyTorch with Python's asyncio, our approach is validated through experiments demonstrating improved performance and scalability for asynchronous, heterogeneous, and non-convex FL scenarios.
Similar Papers
Client-Centric Federated Adaptive Optimization
Machine Learning (CS)
Helps computers learn together without sharing private data.
The Impact Analysis of Delays in Asynchronous Federated Learning with Data Heterogeneity for Edge Intelligence
Machine Learning (CS)
Lets computers learn together even with slow connections.
Mitigating Participation Imbalance Bias in Asynchronous Federated Learning
Machine Learning (CS)
Makes AI learn better from many different computers.