Gradient Descent as a Perceptron Algorithm: Understanding Dynamics and Implicit Acceleration
By: Alexander Tyurin
Even for the gradient descent (GD) method applied to neural network training, understanding its optimization dynamics, including convergence rate, iterate trajectories, function value oscillations, and especially its implicit acceleration, remains a challenging problem. We analyze nonlinear models with the logistic loss and show that the steps of GD reduce to those of generalized perceptron algorithms (Rosenblatt, 1958), providing a new perspective on the dynamics. This reduction yields significantly simpler algorithmic steps, which we analyze using classical linear algebra tools. Using these tools, we demonstrate on a minimalistic example that the nonlinearity in a two-layer model can provably yield a faster iteration complexity $\tilde{O}(\sqrt{d})$ compared to $Ω(d)$ achieved by linear models, where $d$ is the number of features. This helps explain the optimization dynamics and the implicit acceleration phenomenon observed in neural networks. The theoretical results are supported by extensive numerical experiments. We believe that this alternative view will further advance research on the optimization of neural networks.
Similar Papers
On the Theory of Continual Learning with Gradient Descent for Neural Networks
Machine Learning (Stat)
Helps AI remember old lessons while learning new ones.
Non-Singularity of the Gradient Descent map for Neural Networks with Piecewise Analytic Activations
Optimization and Control
Helps computers learn better, even complex ones.
A Local Polyak-Lojasiewicz and Descent Lemma of Gradient Descent For Overparametrized Linear Models
Machine Learning (CS)
Makes computer learning faster with fewer rules.