Adaptive Conditional Gradient Descent
By: Abbas Khademi, Antonio Silveti-Falls
Potential Business Impact:
Makes computer learning faster and better.
Selecting an effective step-size is a fundamental challenge in first-order optimization, especially for problems with non-Euclidean geometries. This paper presents a novel adaptive step-size strategy for optimization algorithms that rely on linear minimization oracles, as used in the Conditional Gradient or non-Euclidean Normalized Steepest Descent algorithms. Using a simple heuristic to estimate a local Lipschitz constant for the gradient, we can determine step-sizes that guarantee sufficient decrease at each iteration. More precisely, we establish convergence guarantees for our proposed Adaptive Conditional Gradient Descent algorithm, which covers as special cases both the classical Conditional Gradient algorithm and non-Euclidean Normalized Steepest Descent algorithms with adaptive step-sizes. Our analysis covers optimization of continuously differentiable functions in non-convex, quasar-convex, and strongly convex settings, achieving convergence rates that match state-of-the-art theoretical bounds. Comprehensive numerical experiments validate our theoretical findings and illustrate the practical effectiveness of Adaptive Conditional Gradient Descent. The results exhibit competitive performance, underscoring the potential of the adaptive step-size for applications.
Similar Papers
Adaptive control mechanisms in gradient descent algorithms
Optimization and Control
Makes computer learning faster and more accurate.
Stochastic Adaptive Gradient Descent Without Descent
Machine Learning (CS)
Makes computer learning faster without needing extra settings.
A Tale of Two Geometries: Adaptive Optimizers and Non-Euclidean Descent
Machine Learning (CS)
Makes computer learning faster and better.