MD-SNN: Membrane Potential-aware Distillation on Quantized Spiking Neural Network
By: Donghyun Lee , Abhishek Moitra , Youngeun Kim and more
Potential Business Impact:
Makes smart computer brains use less power.
Spiking Neural Networks (SNNs) offer a promising and energy-efficient alternative to conventional neural networks, thanks to their sparse binary activation. However, they face challenges regarding memory and computation overhead due to complex spatio-temporal dynamics and the necessity for multiple backpropagation computations across timesteps during training. To mitigate this overhead, compression techniques such as quantization are applied to SNNs. Yet, naively applying quantization to SNNs introduces a mismatch in membrane potential, a crucial factor for the firing of spikes, resulting in accuracy degradation. In this paper, we introduce Membrane-aware Distillation on quantized Spiking Neural Network (MD-SNN), which leverages membrane potential to mitigate discrepancies after weight, membrane potential, and batch normalization quantization. To our knowledge, this study represents the first application of membrane potential knowledge distillation in SNNs. We validate our approach on various datasets, including CIFAR10, CIFAR100, N-Caltech101, and TinyImageNet, demonstrating its effectiveness for both static and dynamic data scenarios. Furthermore, for hardware efficiency, we evaluate the MD-SNN with SpikeSim platform, finding that MD-SNNs achieve 14.85X lower energy-delay-area product (EDAP), 2.64X higher TOPS/W, and 6.19X higher TOPS/mm2 compared to floating point SNNs at iso-accuracy on N-Caltech101 dataset.
Similar Papers
Memory-Free and Parallel Computation for Quantized Spiking Neural Networks
Neural and Evolutionary Computing
Makes AI smarter using less power.
Quantization Meets Spikes: Lossless Conversion in the First Timestep via Polarity Multi-Spike Mapping
Neural and Evolutionary Computing
Makes AI learn faster and use less power.
Efficient Deployment of Spiking Neural Networks on SpiNNaker2 for DVS Gesture Recognition Using Neuromorphic Intermediate Representation
Machine Learning (CS)
Lets robots learn faster using less power.