The Hidden Cost of Approximation in Online Mirror Descent
By: Ofir Schlisselberg , Uri Sherman , Tomer Koren and more
Potential Business Impact:
Makes computer learning more accurate with bad data.
Online mirror descent (OMD) is a fundamental algorithmic paradigm that underlies many algorithms in optimization, machine learning and sequential decision-making. The OMD iterates are defined as solutions to optimization subproblems which, oftentimes, can be solved only approximately, leading to an inexact version of the algorithm. Nonetheless, existing OMD analyses typically assume an idealized error free setting, thereby limiting our understanding of performance guarantees that should be expected in practice. In this work we initiate a systematic study into inexact OMD, and uncover an intricate relation between regularizer smoothness and robustness to approximation errors. When the regularizer is uniformly smooth, we establish a tight bound on the excess regret due to errors. Then, for barrier regularizers over the simplex and its subsets, we identify a sharp separation: negative entropy requires exponentially small errors to avoid linear regret, whereas log-barrier and Tsallis regularizers remain robust even when the errors are only polynomial. Finally, we show that when the losses are stochastic and the domain is the simplex, negative entropy regains robustness-but this property does not extend to all subsets, where exponentially small errors are again necessary to avoid suboptimal regret.
Similar Papers
Variational Online Mirror Descent for Robust Learning in Schrödinger Bridge
Machine Learning (CS)
Makes AI learn better and faster.
Mirror Descent Using the Tempesta Generalized Multi-parametric Logarithms
Machine Learning (Stat)
Makes computer learning smarter and faster.
Early-Stopped Mirror Descent for Linear Regression over Convex Bodies
Machine Learning (CS)
Finds best answers when data is tricky.