Bandit Learning in Housing Markets
By: Shiyun Lin
Potential Business Impact:
Helps people find best homes by learning preferences.
The housing market, also known as one-sided matching market, is a classic exchange economy model where each agent on the demand side initially owns an indivisible good (a house) and has a personal preference over all goods. The goal is to find a core-stable allocation that exhausts all mutually beneficial exchanges among subgroups of agents. While this model has been extensively studied in economics and computer science due to its broad applications, little attention has been paid to settings where preferences are unknown and must be learned through repeated interactions. In this paper, we propose a statistical learning model within the multi-player multi-armed bandit framework, where players (agents) learn their preferences over arms (goods) from stochastic rewards. We introduce the notion of core regret for each player as the market objective. We study both centralized and decentralized approaches, proving $O(N \log T / Δ^2)$ upper bounds on regret, where $N$ is the number of players, $T$ is the time horizon and $Δ$ is the minimum preference gap among players. For the decentralized setting, we also establish a matching lower bound, demonstrating that our algorithm is order-optimal.
Similar Papers
Optimal Algorithms for Bandit Learning in Matching Markets
CS and Game Theory
Finds best job matches faster with less guessing.
Learning Equilibria in Matching Games with Bandit Feedback
Machine Learning (CS)
Helps matching systems learn fair deals for everyone.
Bandits with Single-Peaked Preferences and Limited Resources
Machine Learning (CS)
Helps computers pick the best things for people.