FPGA-based Acceleration for Convolutional Neural Networks: A Comprehensive Review
By: Junye Jiang , Yaan Zhou , Yuanhao Gong and more
Potential Business Impact:
Makes smart computer programs run faster and cheaper.
Convolutional Neural Networks (CNNs) are fundamental to deep learning, driving applications across various domains. However, their growing complexity has significantly increased computational demands, necessitating efficient hardware accelerators. Field-Programmable Gate Arrays (FPGAs) have emerged as a leading solution, offering reconfigurability, parallelism, and energy efficiency. This paper provides a comprehensive review of FPGA-based hardware accelerators specifically designed for CNNs. It presents and summarizes the performance evaluation framework grounded in existing studies and explores key optimization strategies, such as parallel computing, dataflow optimization, and hardware-software co-design. It also compares various FPGA architectures in terms of latency, throughput, compute efficiency, power consumption, and resource utilization. Finally, the paper highlights future challenges and opportunities, emphasizing the potential for continued innovation in this field.
Similar Papers
Real Time FPGA Based CNNs for Detection, Classification, and Tracking in Autonomous Systems: State of the Art Designs and Optimizations
Hardware Architecture
Makes cameras understand things faster and with less power.
A Resource-Driven Approach for Implementing CNNs on FPGAs Using Adaptive IPs
Hardware Architecture
Makes AI run faster on small chips.
Energy-Efficient FPGA Framework for Non-Quantized Convolutional Neural Networks
Hardware Architecture
Makes AI faster and smarter on small devices.