Uncertainty Quantification for Ranking with Heterogeneous Preferences
By: Jianqing Fan, Hyukjun Kwon, Xiaonan Zhu
Potential Business Impact:
Learns what people like from their choices.
This paper studies human preference learning based on partially revealed choice behavior and formulates the problem as a generalized Bradley-Terry-Luce (BTL) ranking model that accounts for heterogeneous preferences. Specifically, we assume that each user is associated with a nonparametric preference function, and each item is characterized by a low-dimensional latent feature vector - their interaction defines the underlying low-rank score matrix. In this formulation, we propose an indirect regularization method for collaboratively learning the score matrix, which ensures entrywise $\ell_\infty$-norm error control - a novel contribution to the heterogeneous preference learning literature. This technique is based on sieve approximation and can be extended to a broader class of binary choice models where a smooth link function is adopted. In addition, by applying a single step of the Newton-Raphson method, we debias the regularized estimator and establish uncertainty quantification for item scores and rankings of items, both for the aggregated and individual preferences. Extensive simulation results from synthetic and real datasets corroborate our theoretical findings.
Similar Papers
Uncertainty Quantification for Large Language Model Reward Learning under Heterogeneous Human Feedback
Machine Learning (Stat)
Makes AI understand what people like better.
Preference-Based Dynamic Ranking Structure Recognition
Machine Learning (Stat)
Find hidden patterns in people's choices.
Learning when to rank: Estimation of partial rankings from sparse, noisy comparisons
Physics and Society
Finds true winners when data is messy.