GRU: Mitigating the Trade-off between Unlearning and Retention for LLMs
By: Yue Wang , Qizhou Wang , Feng Liu and more
Potential Business Impact:
Cleans AI brains without breaking other skills.
Large language model (LLM) unlearning has demonstrated its essential role in removing privacy and copyright-related responses, crucial for their legal and safe applications. However, the pursuit of complete unlearning often comes with substantial costs due to its compromises in their general functionality, leading to a notorious trade-off between unlearning and retention. It motivates this paper to explore enhanced unlearning schemes that can mitigate this trade-off. Specifically, we propose Gradient Rectified Unlearning (GRU), an improved framework that regulates the directions of gradient updates during the unlearning procedure such that their side impacts on other, unrelated responses can be minimized. GRU is easy and general to implement, demonstrating practical effectiveness across a variety of well-established unlearning benchmarks.
Similar Papers
LLM Unlearning using Gradient Ratio-Based Influence Estimation and Noise Injection
Machine Learning (CS)
Removes specific data from AI without breaking it.
GRAIL: Gradient-Based Adaptive Unlearning for Privacy and Copyright in LLMs
Computation and Language
Removes private info from AI without breaking it.
A Survey on Unlearning in Large Language Models
Computation and Language
Lets AI forget private or bad information.