PABBO: Preferential Amortized Black-Box Optimization
By: Xinyu Zhang , Daolang Huang , Samuel Kaski and more
Potential Business Impact:
Teaches computers to learn what people like faster.
Preferential Bayesian Optimization (PBO) is a sample-efficient method to learn latent user utilities from preferential feedback over a pair of designs. It relies on a statistical surrogate model for the latent function, usually a Gaussian process, and an acquisition strategy to select the next candidate pair to get user feedback on. Due to the non-conjugacy of the associated likelihood, every PBO step requires a significant amount of computations with various approximate inference techniques. This computational overhead is incompatible with the way humans interact with computers, hindering the use of PBO in real-world cases. Building on the recent advances of amortized BO, we propose to circumvent this issue by fully amortizing PBO, meta-learning both the surrogate and the acquisition function. Our method comprises a novel transformer neural process architecture, trained using reinforcement learning and tailored auxiliary losses. On a benchmark composed of synthetic and real-world datasets, our method is several orders of magnitude faster than the usual Gaussian process-based strategies and often outperforms them in accuracy.
Similar Papers
Constrained Preferential Bayesian Optimization and Its Application in Banner Ad Design
Machine Learning (CS)
Helps computers design things people like better.
AutoPBO: LLM-powered Optimization for Local Search PBO Solvers
Artificial Intelligence
Makes computer programs solve hard problems faster.
Clustering-based Meta Bayesian Optimization with Theoretical Guarantee
Machine Learning (CS)
Finds best settings faster, even with many past tries.