Coherence-Aware Distributed Learning under Heterogeneous Downlink Impairments
By: Mehdi Karbalayghareh, David J. Love, Christopher G. Brinton
Potential Business Impact:
Helps phones learn together faster wirelessly.
The performance of federated learning (FL) over wireless networks critically depends on accurate and timely channel state information (CSI) across distributed devices. This requirement is tightly linked to how rapidly the channel gains vary, i.e., the coherence intervals. In practice, edge devices often exhibit unequal coherence times due to differences in mobility and scattering environments, leading to unequal demands for pilot signaling and channel estimation resources. Conventional FL schemes that overlook this coherence disparity can suffer from severe communication inefficiencies and training overhead. This paper proposes a coherence-aware, communication-efficient framework for joint channel training and model updating in practical wireless FL systems operating under heterogeneous fading dynamics. Focusing on downlink impairments, we introduce a resource-reuse strategy based on product superposition, enabling the parameter server to efficiently schedule both static and dynamic devices by embedding global model updates for static devices within pilot transmissions intended for mobile devices. We theoretically analyze the convergence behavior of the proposed scheme and quantify its gains in expected communication efficiency and training accuracy. Experiments demonstrate the effectiveness of the proposed framework under mobility-induced dynamics and offer useful insights for the practical deployment of FL over wireless channels.
Similar Papers
Analog Over-the-Air Federated Learning with Interference-Based Energy Harvesting
Information Theory
Makes wireless learning work better with less power.
Joint Channel and Semantic-aware Grouping for Effective Collaborative Edge Inference
Information Theory
Helps devices work together to fix bad data.
Integrated user scheduling and beam steering in over-the-air federated learning for mobile IoT
Distributed, Parallel, and Cluster Computing
Helps phones learn without sharing private data.