Leveraging Neural Graph Compilers in Machine Learning Research for Edge-Cloud Systems
By: Alireza Furutanpey , Carmen Walser , Philipp Raith and more
Potential Business Impact:
Makes AI run faster on different computers.
This work presents a comprehensive evaluation of neural network graph compilers across heterogeneous hardware platforms, addressing the critical gap between theoretical optimization techniques and practical deployment scenarios. We demonstrate how vendor-specific optimizations can invalidate relative performance comparisons between architectural archetypes, with performance advantages sometimes completely reversing after compilation. Our systematic analysis reveals that graph compilers exhibit performance patterns highly dependent on both neural architecture and batch sizes. Through fine-grained block-level experimentation, we establish that vendor-specific compilers can leverage repeated patterns in simple architectures, yielding disproportionate throughput gains as model depth increases. We introduce novel metrics to quantify a compiler's ability to mitigate performance friction as batch size increases. Our methodology bridges the gap between academic research and practical deployment by incorporating compiler effects throughout the research process, providing actionable insights for practitioners navigating complex optimization landscapes across heterogeneous hardware environments.
Similar Papers
Distributed Graph Neural Network Inference With Just-In-Time Compilation For Industry-Scale Graphs
Machine Learning (CS)
Makes big computer graphs learn much faster.
Hardware-Aware Neural Network Compilation with Learned Optimization: A RISC-V Accelerator Approach
Hardware Architecture
Makes computer chips run faster and use less power.
Theoretical Foundations of GPU-Native Compilation for Rapid Code Iteration
Distributed, Parallel, and Cluster Computing
Makes AI code run much, much faster.