Stochastic momentum ADMM for nonconvex and nonsmooth optimization with application to PnP algorithm
By: Kangkang Deng , Shuchang Zhang , Boyu Wang and more
Potential Business Impact:
Solves hard math problems faster and better.
This paper proposes SMADMM, a single-loop Stochastic Momentum Alternating Direction Method of Multipliers for solving a class of nonconvex and nonsmooth composite optimization problems. SMADMM achieves the optimal oracle complexity of $\mathcal{O}(\epsilon^{-3/2})$ in the online setting. Unlike previous stochastic ADMM algorithms that require large mini-batches or a double-loop structure, SMADMM uses only $\mathcal{O}(1)$ stochastic gradient evaluations per iteration and avoids costly restarts. To further improve practicality, we incorporate dynamic step sizes and penalty parameters, proving that SMADMM maintains its optimal complexity without the need for large initial batches. We also develop PnP-SMADMM by integrating plug-and-play priors, and establish its theoretical convergence under mild assumptions. Extensive experiments on classification, CT image reconstruction, and phase retrieval tasks demonstrate that our approach outperforms existing stochastic ADMM methods both in accuracy and efficiency, validating our theoretical results.
Similar Papers
BC-ADMM: An Efficient Non-convex Constrained Optimizer with Robotic Applications
Optimization and Control
Robots move faster and smarter with new math.
Smoothing ADMM for Non-convex and Non-smooth Hierarchical Federated Learning
Machine Learning (CS)
Trains AI smarter and faster with different data.
Stochastic Momentum Methods for Non-smooth Non-Convex Finite-Sum Coupled Compositional Optimization
Machine Learning (CS)
Makes computer learning faster and better.