Learning in Repeated Multi-Objective Stackelberg Games with Payoff Manipulation
By: Phurinut Srisawad, Juergen Branke, Long Tran-Thanh
Potential Business Impact:
Helps one person get more from another.
We study payoff manipulation in repeated multi-objective Stackelberg games, where a leader may strategically influence a follower's deterministic best response, e.g., by offering a share of their own payoff. We assume that the follower's utility function, representing preferences over multiple objectives, is unknown but linear, and its weight parameter must be inferred through interaction. This introduces a sequential decision-making challenge for the leader, who must balance preference elicitation with immediate utility maximisation. We formalise this problem and propose manipulation policies based on expected utility (EU) and long-term expected utility (longEU), which guide the leader in selecting actions and offering incentives that trade off short-term gains with long-term impact. We prove that under infinite repeated interactions, longEU converges to the optimal manipulation. Empirical results across benchmark environments demonstrate that our approach improves cumulative leader utility while promoting mutually beneficial outcomes, all without requiring explicit negotiation or prior knowledge of the follower's utility function.
Similar Papers
Learning in Repeated Multi-Objective Stackelberg Games with Payoff Manipulation
CS and Game Theory
Helps one person get more from another.
Learning Local Stackelberg Equilibria from Repeated Interactions with a Learning Agent
CS and Game Theory
Helps bosses get better deals from smart workers.
Learning in Stackelberg Markov Games
Systems and Control
Helps set fair electricity prices for everyone.