RapidUn: Influence-Driven Parameter Reweighting for Efficient Large Language Model Unlearning
By: Guoshenghui Zhao, Huawei Lin, Weijie Zhao
Potential Business Impact:
Teaches AI to forget bad information quickly.
Removing specific data influence from large language models (LLMs) remains challenging, as retraining is costly and existing approximate unlearning methods are often unstable. The challenge is exacerbated when the forget set is small or imbalanced. We introduce RapidUn, an influence-driven and parameter-efficient unlearning framework. It first estimates per-sample influence through a fast estimation module, then maps these scores into adaptive update weights that guide selective parameter updates -- forgetting harmful behavior while retaining general knowledge. On Mistral-7B and Llama-3-8B across Dolly-15k and Alpaca-57k, RapidUn achieves up to 100 times higher efficiency than full retraining and consistently outperforms Fisher, GA, and LoReUn on both in-distribution and out-of-distribution forgetting. These results establish influence-guided parameter reweighting as a scalable and interpretable paradigm for LLM unlearning.
Similar Papers
Unlearning That Lasts: Utility-Preserving, Robust, and Almost Irreversible Forgetting in LLMs
Machine Learning (CS)
Removes bad info from AI, making it safer.
LLM Unlearning using Gradient Ratio-Based Influence Estimation and Noise Injection
Machine Learning (CS)
Removes specific data from AI without breaking it.
Reveal and Release: Iterative LLM Unlearning with Self-generated Data
Computation and Language
Teaches computers to forget private or bad information.