Score: 2

Vul-R2: A Reasoning LLM for Automated Vulnerability Repair

Published: October 7, 2025 | arXiv ID: 2510.05480v1

By: Xin-Cheng Wen , Zirui Lin , Yijun Yang and more

BigTech Affiliations: Tencent

Potential Business Impact:

Fixes computer bugs automatically using smart programs.

Business Areas:
Natural Language Processing Artificial Intelligence, Data and Analytics, Software

The exponential increase in software vulnerabilities has created an urgent need for automatic vulnerability repair (AVR) solutions. Recent research has formulated AVR as a sequence generation problem and has leveraged large language models (LLMs) to address this problem. Typically, these approaches prompt or fine-tune LLMs to generate repairs for vulnerabilities directly. Although these methods show state-of-the-art performance, they face the following challenges: (1) Lack of high-quality, vulnerability-related reasoning data. Current approaches primarily rely on foundation models that mainly encode general programming knowledge. Without vulnerability-related reasoning data, they tend to fail to capture the diverse vulnerability repair patterns. (2) Hard to verify the intermediate vulnerability repair process during LLM training. Existing reinforcement learning methods often leverage intermediate execution feedback from the environment (e.g., sandbox-based execution results) to guide reinforcement learning training. In contrast, the vulnerability repair process generally lacks such intermediate, verifiable feedback, which poses additional challenges for model training.

Country of Origin
🇨🇳 China

Page Count
13 pages

Category
Computer Science:
Artificial Intelligence