Novel Pivoted Cholesky Decompositions for Efficient Gaussian Process Inference
By: Filip de Roos, Fabio Muratore
Potential Business Impact:
Finds better math shortcuts for computers.
The Cholesky decomposition is a fundamental tool for solving linear systems with symmetric and positive definite matrices which are ubiquitous in linear algebra, optimization, and machine learning. Its numerical stability can be improved by introducing a pivoting strategy that iteratively permutes the rows and columns of the matrix. The order of pivoting indices determines how accurately the intermediate decomposition can reconstruct the original matrix, thus is decisive for the algorithm's efficiency in the case of early termination. Standard implementations select the next pivot from the largest value on the diagonal. In the case of Bayesian nonparametric inference, this strategy corresponds to greedy entropy maximization, which is often used in active learning and design of experiments. We explore this connection in detail and deduce novel pivoting strategies for the Cholesky decomposition. The resulting algorithms are more efficient at reducing the uncertainty over a data set, can be updated to include information about observations, and additionally benefit from a tailored implementation. We benchmark the effectiveness of the new selection strategies on two tasks important to Gaussian processes: sparse regression and inference based on preconditioned iterative solvers. Our results show that the proposed selection strategies are either on par or, in most cases, outperform traditional baselines while requiring a negligible amount of additional computation.
Similar Papers
The Geometry of the Pivot: A Note on Lazy Pivoted Cholesky and Farthest Point Sampling
Machine Learning (CS)
Makes big computer learning faster and easier.
The Geometry of the Pivot: A Note on Lazy Pivoted Cholesky and Farthest Point Sampling
Machine Learning (CS)
Makes big computer learning faster and smarter.
Accelerated decomposition of bistochastic kernel matrices by low rank approximation
Numerical Analysis
Finds hidden patterns in messy data faster.