Score: 0

Energy-convergence trade off for the training of neural networks on bio-inspired hardware

Published: September 10, 2025 | arXiv ID: 2509.18121v1

By: Nikhil Garg , Paul Uriarte Vicandi , Yanming Zhang and more

Potential Business Impact:

Makes brain-like chips learn faster and use less power.

Business Areas:
Semiconductor Hardware, Science and Engineering

The increasing deployment of wearable sensors and implantable devices is shifting AI processing demands to the extreme edge, necessitating ultra-low power for continuous operation. Inspired by the brain, emerging memristive devices promise to accelerate neural network training by eliminating costly data transfers between compute and memory. Though, balancing performance and energy efficiency remains a challenge. We investigate ferroelectric synaptic devices based on HfO2/ZrO2 superlattices and feed their experimentally measured weight updates into hardware-aware neural network simulations. Across pulse widths from 20 ns to 0.2 ms, shorter pulses lower per-update energy but require more training epochs while still reducing total energy without sacrificing accuracy. Classification accuracy using plain stochastic gradient descent (SGD) is diminished compared to mixed-precision SGD. We analyze the causes and propose a ``symmetry point shifting'' technique, addressing asymmetric updates and restoring accuracy. These results highlight a trade-off among accuracy, convergence speed, and energy use, showing that short-pulse programming with tailored training significantly enhances on-chip learning efficiency.

Country of Origin
🇨🇭 Switzerland

Page Count
5 pages

Category
Computer Science:
Emerging Technologies