Differential Coding for Training-Free ANN-to-SNN Conversion
By: Zihan Huang , Wei Fang , Tong Bu and more
Potential Business Impact:
Makes AI use less power and be faster.
Spiking Neural Networks (SNNs) exhibit significant potential due to their low energy consumption. Converting Artificial Neural Networks (ANNs) to SNNs is an efficient way to achieve high-performance SNNs. However, many conversion methods are based on rate coding, which requires numerous spikes and longer time-steps compared to directly trained SNNs, leading to increased energy consumption and latency. This article introduces differential coding for ANN-to-SNN conversion, a novel coding scheme that reduces spike counts and energy consumption by transmitting changes in rate information rather than rates directly, and explores its application across various layers. Additionally, the threshold iteration method is proposed to optimize thresholds based on activation distribution when converting Rectified Linear Units (ReLUs) to spiking neurons. Experimental results on various Convolutional Neural Networks (CNNs) and Transformers demonstrate that the proposed differential coding significantly improves accuracy while reducing energy consumption, particularly when combined with the threshold iteration method, achieving state-of-the-art performance. The source codes of the proposed method are available at https://github.com/h-z-h-cell/ANN-to-SNN-DCGS.
Similar Papers
Efficient ANN-SNN Conversion with Error Compensation Learning
Machine Learning (CS)
Makes smart computer brains work faster on small devices.
Hybrid Layer-Wise ANN-SNN With Surrogate Spike Encoding-Decoding Structure
Neural and Evolutionary Computing
Makes smart computers use less power.
Lightweight LIF-only SNN accelerator using differential time encoding
Neural and Evolutionary Computing
Makes AI learn faster and use less power.