Wasserstein Barycenter Gaussian Process based Bayesian Optimization
By: Antonio Candelieri, Andrea Ponti, Francesco Archetti
Potential Business Impact:
Finds best answers faster, even on hard problems.
Gaussian Process based Bayesian Optimization is a widely applied algorithm to learn and optimize under uncertainty, well-known for its sample efficiency. However, recently -- and more frequently -- research studies have empirically demonstrated that the Gaussian Process fitting procedure at its core could be its most relevant weakness. Fitting a Gaussian Process means tuning its kernel's hyperparameters to a set of observations, but the common Maximum Likelihood Estimation technique, usually appropriate for learning tasks, has shown different criticalities in Bayesian Optimization, making theoretical analysis of this algorithm an open challenge. Exploiting the analogy between Gaussian Processes and Gaussian Distributions, we present a new approach which uses a prefixed set of hyperparameters values to fit as many Gaussian Processes and then combines them into a unique model as a Wasserstein Barycenter of Gaussian Processes. We considered both "easy" test problems and others known to undermine the \textit{vanilla} Bayesian Optimization algorithm. The new method, namely Wasserstein Barycenter Gausssian Process based Bayesian Optimization (WBGP-BO), resulted promising and able to converge to the optimum, contrary to vanilla Bayesian Optimization, also on the most "tricky" test problems.
Similar Papers
Collaborative Bayesian Optimization via Wasserstein Barycenters
Machine Learning (CS)
Helps computers learn secrets without sharing data.
Gradient-based Sample Selection for Faster Bayesian Optimization
Machine Learning (Stat)
Makes computer searches faster by picking smart data.
Constrained Bayesian Optimization under Bivariate Gaussian Process with Application to Cure Process Optimization
Computation
Finds best settings faster, even with tricky rules.