Near-Optimal Decentralized Stochastic Nonconvex Optimization with Heavy-Tailed Noise
By: Menglian Wang, Zhuanghua Liu, Luo Luo
Potential Business Impact:
Makes computer learning work better with messy data.
This paper studies decentralized stochastic nonconvex optimization problem over row-stochastic networks. We consider the heavy-tailed gradient noise which is empirically observed in many popular real-world applications. Specifically, we propose a decentralized normalized stochastic gradient descent with Pull-Diag gradient tracking, which achieves approximate stationary points with the optimal sample complexity and the near-optimal communication complexity. We further follow our framework to study the setting of undirected networks, also achieving the nearly tight upper complexity bounds. Moreover, we conduct empirical studies to show the practical superiority of the proposed methods.
Similar Papers
Decentralized Nonconvex Optimization under Heavy-Tailed Noise: Normalization and Optimal Convergence
Optimization and Control
Helps computers learn better with messy data.
Nonconvex Decentralized Stochastic Bilevel Optimization under Heavy-Tailed Noises
Machine Learning (CS)
Teaches computers to learn better with messy data.
Decentralized Stochastic Nonconvex Optimization under the Relaxed Smoothness
Optimization and Control
Helps many computers work together to solve problems.