GraphNet: A Large-Scale Computational Graph Dataset for Tensor Compiler Research
By: Xinqi Li , Yiqun Liu , Shan Jiang and more
Potential Business Impact:
Makes computer programs run much faster and correctly.
We introduce GraphNet, a dataset of 2.7K real-world deep learning computational graphs with rich metadata, spanning six major task categories across multiple deep learning frameworks. To evaluate tensor compiler performance on these samples, we propose the benchmark metric Speedup Score S(t), which jointly considers runtime speedup and execution correctness under tunable tolerance levels, offering a reliable measure of general optimization capability. Furthermore, we extend S(t) to the Error-aware Speedup Score ES(t), which incorporates error information and helps compiler developers identify key performance bottlenecks. In this report, we benchmark the default tensor compilers, CINN for PaddlePaddle and TorchInductor for PyTorch, on computer vision (CV) and natural language processing (NLP) samples to demonstrate the practicality of GraphNet. The full construction pipeline with graph extraction and compiler evaluation tools is available at https://github.com/PaddlePaddle/GraphNet .
Similar Papers
GraphBench: Next-generation graph learning benchmarking
Machine Learning (CS)
Makes computer learning on networks faster and fairer.
Distributed Graph Neural Network Inference With Just-In-Time Compilation For Industry-Scale Graphs
Machine Learning (CS)
Makes big computer graphs learn much faster.
Leveraging Neural Graph Compilers in Machine Learning Research for Edge-Cloud Systems
Distributed, Parallel, and Cluster Computing
Makes AI run faster on different computers.