Global Convergence of Four-Layer Matrix Factorization under Random Initialization
By: Minrui Luo , Weihang Xu , Xiang Gao and more
Potential Business Impact:
Makes deep computer learning work better.
Gradient descent dynamics on the deep matrix factorization problem is extensively studied as a simplified theoretical model for deep neural networks. Although the convergence theory for two-layer matrix factorization is well-established, no global convergence guarantee for general deep matrix factorization under random initialization has been established to date. To address this gap, we provide a polynomial-time global convergence guarantee for randomly initialized gradient descent on four-layer matrix factorization, given certain conditions on the target matrix and a standard balanced regularization term. Our analysis employs new techniques to show saddle-avoidance properties of gradient decent dynamics, and extends previous theories to characterize the change in eigenvalues of layer weights.
Similar Papers
Global Convergence of Four-Layer Matrix Factorization under Random Initialization
Optimization and Control
Makes deep computer learning work better.
Understanding Incremental Learning with Closed-form Solution to Gradient Flow on Overparamerterized Matrix Factorization
Machine Learning (CS)
Teaches computers to learn things step-by-step.
Global Convergence Analysis of Vanilla Gradient Descent for Asymmetric Matrix Completion
Machine Learning (CS)
Makes computers fill in missing data faster.