SMoFi: Step-wise Momentum Fusion for Split Federated Learning on Heterogeneous Data
By: Mingkun Yang , Ran Zhu , Qing Wang and more
Potential Business Impact:
Makes AI learn faster even with different data.
Split Federated Learning is a system-efficient federated learning paradigm that leverages the rich computing resources at a central server to train model partitions. Data heterogeneity across silos, however, presents a major challenge undermining the convergence speed and accuracy of the global model. This paper introduces Step-wise Momentum Fusion (SMoFi), an effective and lightweight framework that counteracts gradient divergence arising from data heterogeneity by synchronizing the momentum buffers across server-side optimizers. To control gradient divergence over the training process, we design a staleness-aware alignment mechanism that imposes constraints on gradient updates of the server-side submodel at each optimization step. Extensive validations on multiple real-world datasets show that SMoFi consistently improves global model accuracy (up to 7.1%) and convergence speed (up to 10.25$\times$). Furthermore, SMoFi has a greater impact with more clients involved and deeper learning models, making it particularly suitable for model training in resource-constrained contexts.
Similar Papers
SMoFi: Step-wise Momentum Fusion for Split Federated Learning on Heterogeneous Data
Machine Learning (CS)
Improves AI learning from many computers faster.
A Fast and Flat Federated Learning Method via Weighted Momentum and Sharpness-Aware Minimization
Machine Learning (CS)
Helps computers learn faster from many sources.
Accelerating Wireless Distributed Learning via Hybrid Split and Federated Learning Optimization
Machine Learning (CS)
Makes smart devices learn faster together.