Score: 1

FENIX: Enabling In-Network DNN Inference with FPGA-Enhanced Programmable Switches

Published: July 20, 2025 | arXiv ID: 2507.14891v2

By: Xiangyu Gao , Tong Li , Yinchao Zhang and more

Potential Business Impact:

Makes internet traffic analysis super fast and accurate.

Business Areas:
Field-Programmable Gate Array (FPGA) Hardware

Machine learning (ML) is increasingly used in network data planes for advanced traffic analysis. However, existing solutions (such as FlowLens, N3IC, and BoS) still struggle to simultaneously achieve low latency, high throughput, and high accuracy. To address these challenges, we present FENIX, a hybrid in-network ML system that performs feature extraction on programmable switch ASICs and deep neural network inference on FPGAs. FENIX introduces a Data Engine that leverages a probabilistic token bucket algorithm to control the sending rate of feature streams, effectively addressing the throughput gap between programmable switch ASICs and FPGAs. In addition, FENIX designs a Model Engine to enable high-accuracy deep neural network inference in the network, overcoming the difficulty of deploying complex models on resource-constrained switch chips. We implement FENIX on a programmable switch platform that integrates a Tofino ASIC and a ZU19EG FPGA directly and evaluate it on real-world network traffic datasets. Our results show that FENIX achieves microsecond-level inference latency and multi-terabit throughput with low hardware overhead, and delivers over 95\% accuracy on mainstream network traffic classification tasks, outperforming SOTA.

Page Count
16 pages

Category
Computer Science:
Networking and Internet Architecture