Score: 0

Thompson Sampling for Multi-Objective Linear Contextual Bandit

Published: November 30, 2025 | arXiv ID: 2512.00930v1

By: Somangchan Park, Heesang Ann, Min-hwan Oh

Potential Business Impact:

Helps computers make better choices with many goals.

Business Areas:
A/B Testing Data and Analytics

We study the multi-objective linear contextual bandit problem, where multiple possible conflicting objectives must be optimized simultaneously. We propose \texttt{MOL-TS}, the \textit{first} Thompson Sampling algorithm with Pareto regret guarantees for this problem. Unlike standard approaches that compute an empirical Pareto front each round, \texttt{MOL-TS} samples parameters across objectives and efficiently selects an arm from a novel \emph{effective Pareto front}, which accounts for repeated selections over time. Our analysis shows that \texttt{MOL-TS} achieves a worst-case Pareto regret bound of $\widetilde{O}(d^{3/2}\sqrt{T})$, where $d$ is the dimension of the feature vectors, $T$ is the total number of rounds, matching the best known order for randomized linear bandit algorithms for single objective. Empirical results confirm the benefits of our proposed approach, demonstrating improved regret minimization and strong multi-objective performance.

Country of Origin
🇰🇷 Korea, Republic of

Page Count
33 pages

Category
Statistics:
Machine Learning (Stat)