Bayesian optimal experimental design with Wasserstein information criteria
By: Tapio Helin, Youssef Marzouk, Jose Rodrigo Rojo-Garcia
Potential Business Impact:
Finds best ways to learn from experiments.
Bayesian optimal experimental design (OED) provides a principled framework for selecting the most informative observational settings in experiments. With rapid advances in computational power, Bayesian OED has become increasingly feasible for inference problems involving large-scale simulations, attracting growing interest in fields such as inverse problems. In this paper, we introduce a novel design criterion based on the expected Wasserstein-$p$ distance between the prior and posterior distributions. Especially, for $p=2$, this criterion shares key parallels with the widely used expected information gain (EIG), which relies on the Kullback--Leibler divergence instead. First, the Wasserstein-2 criterion admits a closed-form solution for Gaussian regression, a property which can be also leveraged for approximative schemes. Second, it can be interpreted as maximizing the information gain measured by the transport cost incurred when updating the prior to the posterior. Our main contribution is a stability analysis of the Wasserstein-1 criterion, where we provide a rigorous error analysis under perturbations of the prior or likelihood. We partially extend this study also to the Wasserstein-2 criterion. In particular, these results yield error rates when empirical approximations of priors are used. Finally, we demonstrate the computability of the Wasserstein-2 criterion and demonstrate our approximation rates through simulations.
Similar Papers
Optimal Experimental Design Criteria for Data-Consistent Inversion
Methodology
Finds best experiments to learn more with less work.
Gradient-Free Sequential Bayesian Experimental Design via Interacting Particle Systems
Machine Learning (Stat)
Finds best experiments without complex math.
Modelling bounded rational decision-making through Wasserstein constraints
Machine Learning (CS)
Helps robots learn to make better choices.