Uncertainty-Aware Surrogate-based Amortized Bayesian Inference for Computationally Expensive Models
By: Stefania Scheurer , Philipp Reiser , Tim Brünnette and more
Potential Business Impact:
Makes computer guesses better with less work.
Bayesian inference typically relies on a large number of model evaluations to estimate posterior distributions. Established methods like Markov Chain Monte Carlo (MCMC) and Amortized Bayesian Inference (ABI) can become computationally challenging. While ABI enables fast inference after training, generating sufficient training data still requires thousands of model simulations, which is infeasible for expensive models. Surrogate models offer a solution by providing approximate simulations at a lower computational cost, allowing the generation of large data sets for training. However, the introduced approximation errors and uncertainties can lead to overconfident posterior estimates. To address this, we propose Uncertainty-Aware Surrogate-based Amortized Bayesian Inference (UA-SABI) -- a framework that combines surrogate modeling and ABI while explicitly quantifying and propagating surrogate uncertainties through the inference pipeline. Our experiments show that this approach enables reliable, fast, and repeated Bayesian inference for computationally expensive models, even under tight time constraints.
Similar Papers
Improving the Accuracy of Amortized Model Comparison with Self-Consistency
Machine Learning (Stat)
Makes computer models more reliable when guessing.
Modèles de Substitution pour les Modèles à base d'Agents : Enjeux, Méthodes et Applications
Machine Learning (CS)
Makes complex computer models run much faster.
Towards Trustworthy Amortized Bayesian Model Comparison
Machine Learning (Stat)
Helps computers pick the best explanation for data.