Time-varying Mixing Matrix Design for Energy-efficient Decentralized Federated Learning
By: Xusheng Zhang, Tuan Nguyen, Ting He
We consider the design of mixing matrices to minimize the operation cost for decentralized federated learning (DFL) in wireless networks, with focus on minimizing the maximum per-node energy consumption. As a critical hyperparameter for DFL, the mixing matrix controls both the convergence rate and the needs of agent-to-agent communications, and has thus been studied extensively. However, existing designs mostly focused on minimizing the communication time, leaving open the minimization of per-node energy consumption that is critical for energy-constrained devices. This work addresses this gap through a theoretically-justified solution for mixing matrix design that aims at minimizing the maximum per-node energy consumption until convergence, while taking into account the broadcast nature of wireless communications. Based on a novel convergence theorem that allows arbitrarily time-varying mixing matrices, we propose a multi-phase design framework that activates time-varying communication topologies under optimized budgets to trade off the per-iteration energy consumption and the convergence rate while balancing the energy consumption across nodes. Our evaluations based on real data have validated the efficacy of the proposed solution in combining the low energy consumption of sparse mixing matrices and the fast convergence of dense mixing matrices.
Similar Papers
Communication Optimization for Decentralized Learning atop Bandwidth-limited Edge Networks
Networking and Internet Architecture
Makes smart devices learn faster together.
Towards Heterogeneity-Aware and Energy-Efficient Topology Optimization for Decentralized Federated Learning in Edge Environment
Machine Learning (CS)
Helps many devices train AI without sharing private data.
FairEnergy: Contribution-Based Fairness meets Energy Efficiency in Federated Learning
Machine Learning (CS)
Saves phone battery while learning together.