Sensitivity-Aware Mixed-Precision Quantization for ReRAM-based Computing-in-Memory
By: Guan-Cheng Chen , Chieh-Lin Tsai , Pei-Hsuan Tsai and more
Potential Business Impact:
Makes computer chips use less power for AI.
Compute-In-Memory (CIM) systems, particularly those utilizing ReRAM and memristive technologies, offer a promising path toward energy-efficient neural network computation. However, conventional quantization and compression techniques often fail to fully optimize performance and efficiency in these architectures. In this work, we present a structured quantization method that combines sensitivity analysis with mixed-precision strategies to enhance weight storage and computational performance on ReRAM-based CIM systems. Our approach improves ReRAM Crossbar utilization, significantly reducing power consumption, latency, and computational load, while maintaining high accuracy. Experimental results show 86.33% accuracy at 70% compression, alongside a 40% reduction in power consumption, demonstrating the method's effectiveness for power-constrained applications.
Similar Papers
A Time- and Energy-Efficient CNN with Dense Connections on Memristor-Based Chips
Hardware Architecture
Makes AI chips faster and use less power.
Computing-In-Memory Aware Model Adaption For Edge Devices
Hardware Architecture
Makes AI chips faster and smaller.
Reconfigurable Digital RRAM Logic Enables In-Situ Pruning and Learning for Edge AI
Hardware Architecture
Makes AI learn faster and use less power.