STRIDE: Sparse Techniques for Regression in Deep Gaussian Processes
By: Simon Urbainczyk, Aretha L. Teckentrup, Jonas Latz
Potential Business Impact:
Teaches computers to learn from lots of data.
Gaussian processes (GPs) have gained popularity as flexible machine learning models for regression and function approximation with an in-built method for uncertainty quantification. However, GPs suffer when the amount of training data is large or when the underlying function contains multi-scale features that are difficult to represent by a stationary kernel. To address the former, training of GPs with large-scale data is often performed through inducing point approximations (also known as sparse GP regression (GPR)), where the size of the covariance matrices in GPR is reduced considerably through a greedy search on the data set. To aid the latter, deep GPs have gained traction as hierarchical models that resolve multi-scale features by combining multiple GPs. Posterior inference in deep GPs requires a sampling or, more usual, a variational approximation. Variational approximations lead to large-scale stochastic, non-convex optimisation problems and the resulting approximation tends to represent uncertainty incorrectly. In this work, we combine variational learning with MCMC to develop a particle-based expectation-maximisation method to simultaneously find inducing points within the large-scale data (variationally) and accurately train the GPs (sampling-based). The result is a highly efficient and accurate methodology for deep GP training on large-scale data. We test our method on standard benchmark problems.
Similar Papers
Adaptive sparse variational approximations for Gaussian process regression
Statistics Theory
Helps computers learn better by fixing their settings.
Bayesian Bridge Gaussian Process Regression
Methodology
Finds important information in big data faster.
Scalable Gaussian Processes with Latent Kronecker Structure
Machine Learning (CS)
Lets computers learn from huge, messy data.