Score: 0

Near-Optimal Decentralized Stochastic Nonconvex Optimization with Heavy-Tailed Noise

Published: January 16, 2026 | arXiv ID: 2601.11435v1

By: Menglian Wang, Zhuanghua Liu, Luo Luo

Potential Business Impact:

Makes computer learning work better with messy data.

Business Areas:
Darknet Internet Services

This paper studies decentralized stochastic nonconvex optimization problem over row-stochastic networks. We consider the heavy-tailed gradient noise which is empirically observed in many popular real-world applications. Specifically, we propose a decentralized normalized stochastic gradient descent with Pull-Diag gradient tracking, which achieves approximate stationary points with the optimal sample complexity and the near-optimal communication complexity. We further follow our framework to study the setting of undirected networks, also achieving the nearly tight upper complexity bounds. Moreover, we conduct empirical studies to show the practical superiority of the proposed methods.

Page Count
35 pages

Category
Mathematics:
Optimization and Control