Explainable Preference Learning: a Decision Tree-based Surrogate Model for Preferential Bayesian Optimization
By: Nick Leenders , Thomas Quadt , Boris Cule and more
Current Preferential Bayesian Optimization methods rely on Gaussian Processes (GPs) as surrogate models. These models are hard to interpret, struggle with handling categorical data, and are computationally complex, limiting their real-world usability. In this paper, we introduce an inherently interpretable decision tree-based surrogate model capable of handling both categorical and continuous data, and scalable to large datasets. Extensive numerical experiments on eight increasingly spiky optimization functions show that our model outperforms GP-based alternatives on spiky functions and has only marginally lower performance for non-spiky functions. Moreover, we apply our model to the real-world Sushi dataset and show its ability to learn an individual's sushi preferences. Finally, we show some initial work on using historical preference data to speed up the optimization process for new unseen users.
Similar Papers
Consecutive Preferential Bayesian Optimization
Machine Learning (CS)
Makes smart guesses cheaper and more accurate.
Scalable Decision Focused Learning via Online Trainable Surrogates
Machine Learning (CS)
Teaches computers to make better choices faster.
Preference Construction: A Bayesian Interactive Preference Elicitation Framework Based on Monte Carlo Tree Search
Machine Learning (CS)
Learns what you like faster with fewer questions.