Efficient Estimation of Unfactorizable Systematic Uncertainties
By: Alexis Romero, Kyle Cranmer, Daniel Whiteson
Potential Business Impact:
Finds hidden science clues without guessing.
Accurate assessment of systematic uncertainties is an increasingly vital task in physics studies, where large, high-dimensional datasets, like those collected at the Large Hadron Collider, hold the key to new discoveries. Common approaches to assessing systematic uncertainties rely on simplifications, such as assuming that the impact of the various sources of uncertainty factorizes. In this paper, we provide realistic example scenarios in which this assumption fails. We introduce an algorithm that uses Gaussian process regression to estimate the impact of systematic uncertainties \textit{without} assuming factorization. The Gaussian process models are enhanced with derivative information, which increases the accuracy of the regression without increasing the number of samples. In addition, we present a novel sampling strategy based on Bayesian experimental design, which is shown to be more efficient than random and grid sampling in our example scenarios.
Similar Papers
Efficient multi-fidelity Gaussian process regression for noisy outputs and non-nested experimental designs
Applications
Improves computer predictions with less data.
Data-driven Approach for Interpolation of Sparse Data
Data Analysis, Statistics and Probability
Combines messy science data for better understanding.
Uncertainty Estimation using Variance-Gated Distributions
Machine Learning (CS)
Makes AI more sure about its answers.