Graph-Dependent Regret Bounds in Multi-Armed Bandits with Interference
By: Fateme Jamshidi, Mohammad Shahverdikondori, Negar Kiyavash
Potential Business Impact:
Helps smart systems learn faster with neighbors.
We study multi-armed bandits under network interference, where each unit's reward depends on its own treatment and those of its neighbors in a given graph. This induces an exponentially large action space, making standard approaches computationally impractical. We propose a novel algorithm that uses the local graph structure to minimize regret. We derive a graph-dependent upper bound on cumulative regret that improves over prior work. Additionally, we provide the first lower bounds for bandits with arbitrary network interference, where each bound involves a distinct structural property of the graph. These bounds show that for both dense and sparse graphs, our algorithm is nearly optimal, with matching upper and lower bounds up to logarithmic factors. When the interference graph is unknown, a variant of our algorithm is Pareto optimal: no algorithm can uniformly outperform it across all instances. We complement our theoretical results with numerical experiments, showing that our approach outperforms the baseline methods.
Similar Papers
On the optimal regret of collaborative personalized linear bandits
Machine Learning (CS)
Helps many AI agents learn faster together.
Distributed Multi-Agent Bandits Over Erdős-Rényi Random Networks
Machine Learning (CS)
Helps robots learn faster with less talking.
Stochastic Multi-Objective Multi-Armed Bandits: Regret Definition and Algorithm
Machine Learning (CS)
Helps computers choose best options with many goals.