Energy-convergence trade off for the training of neural networks on bio-inspired hardware
By: Nikhil Garg , Paul Uriarte Vicandi , Yanming Zhang and more
Potential Business Impact:
Makes brain-like chips learn faster and use less power.
The increasing deployment of wearable sensors and implantable devices is shifting AI processing demands to the extreme edge, necessitating ultra-low power for continuous operation. Inspired by the brain, emerging memristive devices promise to accelerate neural network training by eliminating costly data transfers between compute and memory. Though, balancing performance and energy efficiency remains a challenge. We investigate ferroelectric synaptic devices based on HfO2/ZrO2 superlattices and feed their experimentally measured weight updates into hardware-aware neural network simulations. Across pulse widths from 20 ns to 0.2 ms, shorter pulses lower per-update energy but require more training epochs while still reducing total energy without sacrificing accuracy. Classification accuracy using plain stochastic gradient descent (SGD) is diminished compared to mixed-precision SGD. We analyze the causes and propose a ``symmetry point shifting'' technique, addressing asymmetric updates and restoring accuracy. These results highlight a trade-off among accuracy, convergence speed, and energy use, showing that short-pulse programming with tailored training significantly enhances on-chip learning efficiency.
Similar Papers
Unsupervised local learning based on voltage-dependent synaptic plasticity for resistive and ferroelectric synapses
Neural and Evolutionary Computing
Makes smart gadgets learn with less power.
Enabling Efficient Processing of Spiking Neural Networks with On-Chip Learning on Commodity Neuromorphic Processors for Edge AI Systems
Neural and Evolutionary Computing
Makes robots smarter using less power.
Hardware-aware vs. Hardware-agnostic Energy Estimation for SNN in Space Applications
CV and Pattern Recognition
Saves energy for AI on special chips.