Late Breaking Results: Conversion of Neural Networks into Logic Flows for Edge Computing
By: Daniel Stein , Shaoyi Huang , Rolf Drechsler and more
Potential Business Impact:
Makes smart computer programs run much faster.
Neural networks have been successfully applied in various resource-constrained edge devices, where usually central processing units (CPUs) instead of graphics processing units exist due to limited power availability. State-of-the-art research still focuses on efficiently executing enormous numbers of multiply-accumulate (MAC) operations. However, CPUs themselves are not good at executing such mathematical operations on a large scale, since they are more suited to execute control flow logic, i.e., computer algorithms. To enhance the computation efficiency of neural networks on CPUs, in this paper, we propose to convert them into logic flows for execution. Specifically, neural networks are first converted into equivalent decision trees, from which decision paths with constant leaves are then selected and compressed into logic flows. Such logic flows consist of if and else structures and a reduced number of MAC operations. Experimental results demonstrate that the latency can be reduced by up to 14.9 % on a simulated RISC-V CPU without any accuracy degradation. The code is open source at https://github.com/TUDa-HWAI/NN2Logic
Similar Papers
Dedicated Inference Engine and Binary-Weight Neural Networks for Lightweight Instance Segmentation
CV and Pattern Recognition
Makes smart computer chips use less power and space.
Biological Intuition on Digital Hardware: An RTL Implementation of Poisson-Encoded SNNs for Static Image Classification
Hardware Architecture
Makes smart devices use less power for thinking.
Continuous-Flow Data-Rate-Aware CNN Inference on FPGA
Machine Learning (CS)
Makes AI chips work faster and use less power.