In-memory Training on Analog Devices with Limited Conductance States via Multi-tile Residual Learning
By: Jindan Li , Zhaoxian Wu , Gaowen Liu and more
Potential Business Impact:
Trains AI better with cheaper, simpler computer parts.
Analog in-memory computing (AIMC) accelerators enable efficient deep neural network computation directly within memory using resistive crossbar arrays, where model parameters are represented by the conductance states of memristive devices. However, effective in-memory training typically requires at least 8-bit conductance states to match digital baselines. Realizing such fine-grained states is costly and often requires complex noise mitigation techniques that increase circuit complexity and energy consumption. In practice, many promising memristive devices such as ReRAM offer only about 4-bit resolution due to fabrication constraints, and this limited update precision substantially degrades training accuracy. To enable on-chip training with these limited-state devices, this paper proposes a \emph{residual learning} framework that sequentially learns on multiple crossbar tiles to compensate the residual errors from low-precision weight updates. Our theoretical analysis shows that the optimality gap shrinks with the number of tiles and achieves a linear convergence rate. Experiments on standard image classification benchmarks demonstrate that our method consistently outperforms state-of-the-art in-memory analog training strategies under limited-state settings, while incurring only moderate hardware overhead as confirmed by our cost analysis.
Similar Papers
Compute SNR-Optimal Analog-to-Digital Converters for Analog In-Memory Computing
Signal Processing
Makes AI faster and use less power.
Guidance and Control Neural Network Acceleration using Memristors
Hardware Architecture
Makes space computers smarter with less power.
A Time- and Energy-Efficient CNN with Dense Connections on Memristor-Based Chips
Hardware Architecture
Makes AI chips faster and use less power.