Learning Local Stackelberg Equilibria from Repeated Interactions with a Learning Agent
By: Nivasini Ananthakrishnan, Yuval Dagan, Kunhe Yang
Potential Business Impact:
Helps bosses get better deals from smart workers.
Motivated by the question of how a principal can maximize its utility in repeated interactions with a learning agent, we study repeated games between an principal and an agent employing a mean-based learning algorithm. Prior work has shown that computing or even approximating the global Stackelberg value in similar settings can require an exponential number of rounds in the size of the agent's action space, making it computationally intractable. In contrast, we shift focus to the computation of local Stackelberg equilibria and introduce an algorithm that, within the smoothed analysis framework, constitutes a Polynomial Time Approximation Scheme (PTAS) for finding an epsilon-approximate local Stackelberg equilibrium. Notably, the algorithm's runtime is polynomial in the size of the agent's action space yet exponential in (1/epsilon) - a dependency we prove to be unavoidable.
Similar Papers
Learning in Stackelberg Markov Games
Systems and Control
Helps set fair electricity prices for everyone.
Learning in Repeated Multi-Objective Stackelberg Games with Payoff Manipulation
CS and Game Theory
Helps one person get more from another.
Learning in Repeated Multi-Objective Stackelberg Games with Payoff Manipulation
CS and Game Theory
Helps one person get more from another.