The global convergence time of stochastic gradient descent in non-convex landscapes: Sharp estimates via large deviations
By: Waïss Azizian , Franck Iutzeler , Jérôme Malick and more
Potential Business Impact:
Helps computers learn faster by finding best answers.
In this paper, we examine the time it takes for stochastic gradient descent (SGD) to reach the global minimum of a general, non-convex loss function. We approach this question through the lens of randomly perturbed dynamical systems and large deviations theory, and we provide a tight characterization of the global convergence time of SGD via matching upper and lower bounds. These bounds are dominated by the most "costly" set of obstacles that the algorithm may need to overcome in order to reach a global minimizer from a given initialization, coupling in this way the global geometry of the underlying loss landscape with the statistics of the noise entering the process. Finally, motivated by applications to the training of deep neural networks, we also provide a series of refinements and extensions of our analysis for loss functions with shallow local minima.
Similar Papers
Quantitative Convergence Analysis of Projected Stochastic Gradient Descent for Non-Convex Losses via the Goldstein Subdifferential
Optimization and Control
Makes AI learn faster without needing extra tricks.
Stochastic Gradient Descent in Non-Convex Problems: Asymptotic Convergence with Relaxed Step-Size via Stopping Time Methods
Machine Learning (CS)
Makes computer learning work with simpler rules.
Long-time dynamics and universality of nonconvex gradient descent
Machine Learning (CS)
Helps computers learn better, even with messy data.