Score: 0

PAC-Bayes Meets Online Contextual Optimization

Published: November 25, 2025 | arXiv ID: 2511.20413v1

By: Zhuojun Xie, Adam Abdin, Yiping Fang

Potential Business Impact:

Helps computers make better choices when things change.

Business Areas:
Semantic Search Internet Services

The predict-then-optimize paradigm bridges online learning and contextual optimization in dynamic environments. Previous works have investigated the sequential updating of predictors using feedback from downstream decisions to minimize regret in the full-information settings. However, existing approaches are predominantly frequentist, rely heavily on gradient-based strategies, and employ deterministic predictors that could yield high variance in practice despite their asymptotic guarantees. This work introduces, to the best of our knowledge, the first Bayesian online contextual optimization framework. Grounded in PAC-Bayes theory and general Bayesian updating principles, our framework achieves $\mathcal{O}(\sqrt{T})$ regret for bounded and mixable losses via a Gibbs posterior, eliminates the dependence on gradients through sequential Monte Carlo samplers, and thereby accommodates nondifferentiable problems. Theoretical developments and numerical experiments substantiate our claims.

Country of Origin
🇫🇷 France

Page Count
15 pages

Category
Mathematics:
Optimization and Control