Graph-based Gossiping for Communication Efficiency in Decentralized Federated Learning
By: Huong Nguyen , Hong-Tri Nguyen , Praveen Kumar Donta and more
Potential Business Impact:
Makes computers learn together faster, safer.
Federated learning has emerged as a privacy-preserving technique for collaborative model training across heterogeneously distributed silos. Yet, its reliance on a single central server introduces potential bottlenecks and risks of single-point failure. Decentralizing the server, often referred to as decentralized learning, addresses this problem by distributing the server role across nodes within the network. One drawback regarding this pure decentralization is it introduces communication inefficiencies, which arise from increased message exchanges in large-scale setups. However, existing proposed solutions often fail to simulate the real-world distributed and decentralized environment in their experiments, leading to unreliable performance evaluations and limited applicability in practice. Recognizing the lack from prior works, this work investigates the correlation between model size and network latency, a critical factor in optimizing decentralized learning communication. We propose a graph-based gossiping mechanism, where specifically, minimum spanning tree and graph coloring are used to optimize network structure and scheduling for efficient communication across various network topologies and message capacities. Our approach configures and manages subnetworks on real physical routers and devices and closely models real-world distributed setups. Experimental results demonstrate that our method significantly improves communication, compatible with different topologies and data sizes, reducing bandwidth and transfer time by up to circa 8 and 4.4 times, respectively, compared to naive flooding broadcasting methods.
Similar Papers
Communication Optimization for Decentralized Learning atop Bandwidth-limited Edge Networks
Networking and Internet Architecture
Makes smart devices learn faster together.
A Gossip-Enhanced Communication Substrate for Agentic AI: Toward Decentralized Coordination in Large-Scale Multi-Agent Systems
Multiagent Systems
Lets many computer agents work together better.
Decentralized Personalization for Federated Medical Image Segmentation via Gossip Contrastive Mutual Learning
Distributed, Parallel, and Cluster Computing
Helps hospitals train AI without sharing patient data.