Sub-microsecond Transformers for Jet Tagging on FPGAs
By: Lauri Laatu , Chang Sun , Arianna Cox and more
Potential Business Impact:
Lets particle colliders see tiny events instantly.
We present the first sub-microsecond transformer implementation on an FPGA achieving competitive performance for state-of-the-art high-energy physics benchmarks. Transformers have shown exceptional performance on multiple tasks in modern machine learning applications, including jet tagging at the CERN Large Hadron Collider (LHC). However, their computational complexity prohibits use in real-time applications, such as the hardware trigger system of the collider experiments up until now. In this work, we demonstrate the first application of transformers for jet tagging on FPGAs, achieving $\mathcal{O}(100)$ nanosecond latency with superior performance compared to alternative baseline models. We leverage high-granularity quantization and distributed arithmetic optimization to fit the entire transformer model on a single FPGA, achieving the required throughput and latency. Furthermore, we add multi-head attention and linear attention support to hls4ml, making our work accessible to the broader fast machine learning community. This work advances the next-generation trigger systems for the High Luminosity LHC, enabling the use of transformers for real-time applications in high-energy physics and beyond.
Similar Papers
TrackCore-F: Deploying Transformer-Based Subatomic Particle Tracking on FPGAs
High Energy Physics - Experiment
Makes AI models run faster on special chips.
Fast Jet Tagging with MLP-Mixers on FPGAs
Instrumentation and Detectors
Faster, smaller computer chips identify particle collisions.
Real Time FPGA Based Transformers & VLMs for Vision Tasks: SOTA Designs and Optimizations
Hardware Architecture
Makes smart AI run faster on small devices.