Scalable and Interpretable Scientific Discovery via Sparse Variational Gaussian Process Kolmogorov-Arnold Networks (SVGP KAN)
By: Y. Sungtaek Ju
Potential Business Impact:
Lets AI learn from more data, understand its guesses.
Kolmogorov-Arnold Networks (KANs) offer a promising alternative to Multi-Layer Perceptron (MLP) by placing learnable univariate functions on network edges, enhancing interpretability. However, standard KANs lack probabilistic outputs, limiting their utility in applications requiring uncertainty quantification. While recent Gaussian Process (GP) extensions to KANs address this, they utilize exact inference methods that scale cubically with data size N, restricting their application to smaller datasets. We introduce the Sparse Variational GP-KAN (SVGP-KAN), an architecture that integrates sparse variational inference with the KAN topology. By employing $M$ inducing points and analytic moment matching, our method reduces computational complexity from $O(N^3)$ to $O(NM^2)$ or linear in sample size, enabling the application of probabilistic KANs to larger scientific datasets. Furthermore, we demonstrate that integrating a permutation-based importance analysis enables the network to function as a framework for structural identification, identifying relevant inputs and classifying functional relationships.
Similar Papers
Transferring Social Network Knowledge from Multiple GNN Teachers to Kolmogorov-Arnold Networks
Machine Learning (CS)
Makes AI learn faster without needing all the connections.
A Primer on Kolmogorov-Arnold Networks (KANs) for Probabilistic Time Series Forecasting
Machine Learning (CS)
Predicts future traffic with less guessing.
Projective Kolmogorov Arnold Neural Networks (P-KANs): Entropy-Driven Functional Space Discovery for Interpretable Machine Learning
Neural and Evolutionary Computing
Makes AI learn better and use fewer computer parts.