Score: 1

JPU: Bridging Jailbreak Defense and Unlearning via On-Policy Path Rectification

Published: January 6, 2026 | arXiv ID: 2601.03005v1

By: Xi Wang , Songlei Jian , Shasha Li and more

Potential Business Impact:

Fixes AI so it can't be tricked into saying bad things.

Business Areas:
Corrections Facilities Privacy and Security

Despite extensive safety alignment, Large Language Models (LLMs) often fail against jailbreak attacks. While machine unlearning has emerged as a promising defense by erasing specific harmful parameters, current methods remain vulnerable to diverse jailbreaks. We first conduct an empirical study and discover that this failure mechanism is caused by jailbreaks primarily activating non-erased parameters in the intermediate layers. Further, by probing the underlying mechanism through which these circumvented parameters reassemble into the prohibited output, we verify the persistent existence of dynamic $\textbf{jailbreak paths}$ and show that the inability to rectify them constitutes the fundamental gap in existing unlearning defenses. To bridge this gap, we propose $\textbf{J}$ailbreak $\textbf{P}$ath $\textbf{U}$nlearning (JPU), which is the first to rectify dynamic jailbreak paths towards safety anchors by dynamically mining on-policy adversarial samples to expose vulnerabilities and identify jailbreak paths. Extensive experiments demonstrate that JPU significantly enhances jailbreak resistance against dynamic attacks while preserving the model's utility.

Country of Origin
πŸ‡¨πŸ‡³ China

Repos / Data Links

Page Count
14 pages

Category
Computer Science:
Cryptography and Security