PAC-Bayes Meets Online Contextual Optimization
By: Zhuojun Xie, Adam Abdin, Yiping Fang
Potential Business Impact:
Helps computers make better choices when things change.
The predict-then-optimize paradigm bridges online learning and contextual optimization in dynamic environments. Previous works have investigated the sequential updating of predictors using feedback from downstream decisions to minimize regret in the full-information settings. However, existing approaches are predominantly frequentist, rely heavily on gradient-based strategies, and employ deterministic predictors that could yield high variance in practice despite their asymptotic guarantees. This work introduces, to the best of our knowledge, the first Bayesian online contextual optimization framework. Grounded in PAC-Bayes theory and general Bayesian updating principles, our framework achieves $\mathcal{O}(\sqrt{T})$ regret for bounded and mixable losses via a Gibbs posterior, eliminates the dependence on gradients through sequential Monte Carlo samplers, and thereby accommodates nondifferentiable problems. Theoretical developments and numerical experiments substantiate our claims.
Similar Papers
Contextual Strongly Convex Simulation Optimization: Optimize then Predict with Inexact Solutions
Machine Learning (Stat)
Helps computers make better choices faster.
Strategy-robust Online Learning in Contextual Pricing
CS and Game Theory
Sells things online even when buyers lie.
Adaptive, Robust and Scalable Bayesian Filtering for Online Learning
Machine Learning (Stat)
Teaches computers to learn and adapt better.