Optimized Architectures for Kolmogorov-Arnold Networks
By: James Bagrow, Josh Bongard
Efforts to improve Kolmogorov-Arnold networks (KANs) with architectural enhancements have been stymied by the complexity those enhancements bring, undermining the interpretability that makes KANs attractive in the first place. Here we study overprovisioned architectures combined with sparsification to learn compact, interpretable KANs without sacrificing accuracy. Crucially, we focus on differentiable sparsification, turning architecture search into an end-to-end optimization problem. Across function approximation benchmarks, dynamical systems forecasting, and real-world prediction tasks, we demonstrate competitive or superior accuracy while discovering substantially smaller models. Overprovisioning and sparsification are synergistic, with the combination outperforming either alone. The result is a principled path toward models that are both more expressive and more interpretable, addressing a key tension in scientific machine learning.
Similar Papers
A Practitioner's Guide to Kolmogorov-Arnold Networks
Machine Learning (CS)
Makes AI smarter and learn faster.
A Practitioner's Guide to Kolmogorov-Arnold Networks
Machine Learning (CS)
Makes computer learning smarter and easier to understand.
Softly Symbolifying Kolmogorov-Arnold Networks
Machine Learning (CS)
Makes AI understand and explain its own thinking.