Sparse Polyak with optimal thresholding operators for high-dimensional M-estimation
By: Tianqi Qiao, Marie Maros
Potential Business Impact:
Finds hidden patterns in big data accurately.
We propose and analyze a variant of Sparse Polyak for high dimensional M-estimation problems. Sparse Polyak proposes a novel adaptive step-size rule tailored to suitably estimate the problem's curvature in the high-dimensional setting, guaranteeing that the algorithm's performance does not deteriorate when the ambient dimension increases. However, convergence guarantees can only be obtained by sacrificing solution sparsity and statistical accuracy. In this work, we introduce a variant of Sparse Polyak that retains its desirable scaling properties with respect to the ambient dimension while obtaining sparser and more accurate solutions.
Similar Papers
Sparse Polyak: an adaptive step size rule for high-dimensional M-estimation
Optimization and Control
Solves hard math problems faster with fewer steps.
High-dimensional limit theorems for SGD: Momentum and Adaptive Step-sizes
Machine Learning (Stat)
Improves computer learning by making it more stable.
Analysis of an Idealized Stochastic Polyak Method and its Application to Black-Box Model Distillation
Machine Learning (CS)
Makes AI learn faster and smaller.