DOSA: Differentiable Model-Based One-Loop Search for DNN Accelerators
By: Charles Hong , Qijing Huang , Grace Dinh and more
Potential Business Impact:
Designs computer chips faster and better.
In the hardware design space exploration process, it is critical to optimize both hardware parameters and algorithm-to-hardware mappings. Previous work has largely approached this simultaneous optimization problem by separately exploring the hardware design space and the mapspace - both individually large and highly nonconvex spaces - independently. The resulting combinatorial explosion has created significant difficulties for optimizers. In this paper, we introduce DOSA, which consists of differentiable performance models and a gradient descent-based optimization technique to simultaneously explore both spaces and identify high-performing design points. Experimental results demonstrate that DOSA outperforms random search and Bayesian optimization by 2.80x and 12.59x, respectively, in improving DNN model energy-delay product, given a similar number of samples. We also demonstrate the modularity and flexibility of DOSA by augmenting our analytical model with a learned model, allowing us to optimize buffer sizes and mappings of a real DNN accelerator and attain a 1.82x improvement in energy-delay product.
Similar Papers
DiffAxE: Diffusion-driven Hardware Accelerator Generation and Design Space Exploration
Hardware Architecture
Finds best computer chips for AI faster.
Comprehensive Design Space Exploration for Tensorized Neural Network Hardware Accelerators
Hardware Architecture
Makes AI run much faster on small devices.
Comprehensive Design Space Exploration for Tensorized Neural Network Hardware Accelerators
Hardware Architecture
Makes smart devices run faster and use less power.