RePro: Leveraging Large Language Models for Semi-Automated Reproduction of Networking Research Results
By: Yining Jiang , Wenyun Xu , Qingyu Song and more
Potential Business Impact:
Helps computers rebuild network programs from papers.
Reproducing networking research is a critical but challenging task due to the scarcity of open-source code. While Large Language Models (LLMs) can automate code generation, current approaches lack the generalizability required for the diverse networking field. To address this, we propose RePro, a semi-automated reproduction framework that leverages advanced prompt engineering to reproduce network systems from their research papers. RePro combines few-shot in-context learning with Structured and Semantic Chain of Thought (SCoT/SeCoT) techniques to systematically translate a paper's description into an optimized, executable implementation. The framework operates through a three-stage pipeline: system description extraction, structural code generation, and code optimization. Our evaluation with five state-of-the-art LLMs across diverse network sub-domains demonstrates that RePro significantly reduces reproduction time compared to manual efforts while achieving comparable system performance, validating its effectiveness and efficiency.
Similar Papers
RePro: Training Language Models to Faithfully Recycle the Web for Pretraining
Computation and Language
Makes AI smarter with less data.
SciReplicate-Bench: Benchmarking LLMs in Agent-driven Algorithmic Reproduction from Research Papers
Computation and Language
Helps computers write code from science papers.
Reflective Paper-to-Code Reproduction Enabled by Fine-Grained Verification
Software Engineering
Helps computers copy science papers into working code.