FedWCM: Unleashing the Potential of Momentum-based Federated Learning in Long-Tailed Scenarios
By: Tianle Li , Yongzhi Huang , Linshan Jiang and more
Potential Business Impact:
Fixes AI learning from messy, uneven data.
Federated Learning (FL) enables decentralized model training while preserving data privacy. Despite its benefits, FL faces challenges with non-identically distributed (non-IID) data, especially in long-tailed scenarios with imbalanced class samples. Momentum-based FL methods, often used to accelerate FL convergence, struggle with these distributions, resulting in biased models and making FL hard to converge. To understand this challenge, we conduct extensive investigations into this phenomenon, accompanied by a layer-wise analysis of neural network behavior. Based on these insights, we propose FedWCM, a method that dynamically adjusts momentum using global and per-round data to correct directional biases introduced by long-tailed distributions. Extensive experiments show that FedWCM resolves non-convergence issues and outperforms existing methods, enhancing FL's efficiency and effectiveness in handling client heterogeneity and data imbalance.
Similar Papers
A Fast and Flat Federated Learning Method via Weighted Momentum and Sharpness-Aware Minimization
Machine Learning (CS)
Helps computers learn faster from many sources.
FedSM: Robust Semantics-Guided Feature Mixup for Bias Reduction in Federated Learning with Long-Tail Data
Machine Learning (CS)
Fixes AI mistakes when data is uneven.
Federated Learning for Financial Forecasting
Machine Learning (CS)
Helps banks predict money changes safely.