KANELÉ: Kolmogorov-Arnold Networks for Efficient LUT-based Evaluation
By: Duc Hoang, Aarush Gupta, Philip Harris
Potential Business Impact:
Makes smart chips learn faster and use less power.
Low-latency, resource-efficient neural network inference on FPGAs is essential for applications demanding real-time capability and low power. Lookup table (LUT)-based neural networks are a common solution, combining strong representational power with efficient FPGA implementation. In this work, we introduce KANELÉ, a framework that exploits the unique properties of Kolmogorov-Arnold Networks (KANs) for FPGA deployment. Unlike traditional multilayer perceptrons (MLPs), KANs employ learnable one-dimensional splines with fixed domains as edge activations, a structure naturally suited to discretization and efficient LUT mapping. We present the first systematic design flow for implementing KANs on FPGAs, co-optimizing training with quantization and pruning to enable compact, high-throughput, and low-latency KAN architectures. Our results demonstrate up to a 2700x speedup and orders of magnitude resource savings compared to prior KAN-on-FPGA approaches. Moreover, KANELÉ matches or surpasses other LUT-based architectures on widely used benchmarks, particularly for tasks involving symbolic or physical formulas, while balancing resource usage across FPGA hardware. Finally, we showcase the versatility of the framework by extending it to real-time, power-efficient control systems.
Similar Papers
Optimizing Neural Networks with Learnable Non-Linear Activation Functions via Lookup-Based FPGA Acceleration
Hardware Architecture
Makes smart devices run faster and use less power.
Hardware Acceleration of Kolmogorov-Arnold Network (KAN) in Large-Scale Systems
Hardware Architecture
Makes AI smarter with less computer parts.
QuantKAN: A Unified Quantization Framework for Kolmogorov Arnold Networks
Machine Learning (CS)
Makes smart computer brains smaller and faster.