Merging of Kolmogorov-Arnold networks trained on disjoint datasets
By: Andrew Polar, Michael Poluektov
Training on disjoint datasets can serve two primary goals: accelerating data processing and enabling federated learning. It has already been established that Kolmogorov-Arnold networks (KANs) are particularly well suited for federated learning and can be merged through simple parameter averaging. While the federated learning literature has mostly focused on achieving training convergence across distributed nodes, the present paper specifically targets acceleration of the training, which depends critically on the choice of an optimisation method and the type of the basis functions. To the best knowledge of the authors, the fastest currently-available combination is the Newton-Kaczmarz method and the piecewise-linear basis functions. Here, it is shown that training on disjoint datasets (or disjoint subsets of the training dataset) can further improve the performance. Experimental comparisons are provided, and all corresponding codes are publicly available.
Similar Papers
Enhancing Federated Learning with Kolmogorov-Arnold Networks: A Comparative Study Across Diverse Aggregation Strategies
Machine Learning (CS)
Makes AI learn better from many computers.
A Practitioner's Guide to Kolmogorov-Arnold Networks
Machine Learning (CS)
Makes computer learning smarter and easier to understand.
A Practitioner's Guide to Kolmogorov-Arnold Networks
Machine Learning (CS)
Makes AI smarter and learn faster.