Score: 2

Global Convergence of Four-Layer Matrix Factorization under Random Initialization

Published: November 13, 2025 | arXiv ID: 2511.09925v1

By: Minrui Luo , Weihang Xu , Xiang Gao and more

BigTech Affiliations: University of Washington

Potential Business Impact:

Makes deep computer learning work better.

Business Areas:
A/B Testing Data and Analytics

Gradient descent dynamics on the deep matrix factorization problem is extensively studied as a simplified theoretical model for deep neural networks. Although the convergence theory for two-layer matrix factorization is well-established, no global convergence guarantee for general deep matrix factorization under random initialization has been established to date. To address this gap, we provide a polynomial-time global convergence guarantee for randomly initialized gradient descent on four-layer matrix factorization, given certain conditions on the target matrix and a standard balanced regularization term. Our analysis employs new techniques to show saddle-avoidance properties of gradient decent dynamics, and extends previous theories to characterize the change in eigenvalues of layer weights.

Country of Origin
πŸ‡ΊπŸ‡Έ πŸ‡¨πŸ‡³ United States, China

Page Count
77 pages

Category
Mathematics:
Optimization and Control