Scalable Gaussian Processes with Latent Kronecker Structure
By: Jihao Andreas Lin , Sebastian Ament , Maximilian Balandat and more
Potential Business Impact:
Lets computers learn from huge, messy data.
Applying Gaussian processes (GPs) to very large datasets remains a challenge due to limited computational scalability. Matrix structures, such as the Kronecker product, can accelerate operations significantly, but their application commonly entails approximations or unrealistic assumptions. In particular, the most common path to creating a Kronecker-structured kernel matrix is by evaluating a product kernel on gridded inputs that can be expressed as a Cartesian product. However, this structure is lost if any observation is missing, breaking the Cartesian product structure, which frequently occurs in real-world data such as time series. To address this limitation, we propose leveraging latent Kronecker structure, by expressing the kernel matrix of observed values as the projection of a latent Kronecker product. In combination with iterative linear system solvers and pathwise conditioning, our method facilitates inference of exact GPs while requiring substantially fewer computational resources than standard iterative methods. We demonstrate that our method outperforms state-of-the-art sparse and variational GPs on real-world datasets with up to five million examples, including robotics, automated machine learning, and climate applications.
Similar Papers
Learning Kronecker-Structured Graphs from Smooth Signals
Machine Learning (CS)
Finds hidden patterns in connected data.
STRIDE: Sparse Techniques for Regression in Deep Gaussian Processes
Machine Learning (Stat)
Teaches computers to learn from lots of data.
Preconditioned Additive Gaussian Processes with Fourier Acceleration
Machine Learning (CS)
Makes computer predictions faster and more accurate.