FeNN-DMA: A RISC-V SoC for SNN acceleration
By: Zainab Aizaz, James C. Knight, Thomas Nowotny
Potential Business Impact:
Makes smart computer brains work faster and use less power.
Spiking Neural Networks (SNNs) are a promising, energy-efficient alternative to standard Artificial Neural Networks (ANNs) and are particularly well-suited to spatio-temporal tasks such as keyword spotting and video classification. However, SNNs have a much lower arithmetic intensity than ANNs and are therefore not well-matched to standard accelerators like GPUs and TPUs. Field Programmable Gate Arrays(FPGAs) are designed for such memory-bound workloads and here we develop a novel, fully-programmable RISC-V-based system-on-chip (FeNN-DMA), tailored to simulating SNNs on modern UltraScale+ FPGAs. We show that FeNN-DMA has comparable resource usage and energy requirements to state-of-the-art fixed-function SNN accelerators, yet it is capable of simulating much larger and more complex models. Using this functionality, we demonstrate state-of-the-art classification accuracy on the Spiking Heidelberg Digits and Neuromorphic MNIST tasks.
Similar Papers
FeNN: A RISC-V vector processor for Spiking Neural Network acceleration
Neural and Evolutionary Computing
Makes AI use much less power.
A Robust, Open-Source Framework for Spiking Neural Networks on Low-End FPGAs
Neural and Evolutionary Computing
Makes brain-like computers run faster on cheap chips.
A Scalable FPGA Architecture With Adaptive Memory Utilization for GEMM-Based Operations
Hardware Architecture
Makes AI learn faster and use less power.