Module-Aware Parameter-Efficient Machine Unlearning on Transformers
By: Wenjie Bao , Jian Lou , Yuke Hu and more
Potential Business Impact:
Removes unwanted data from AI without breaking it.
Transformer has become fundamental to a vast series of pre-trained large models that have achieved remarkable success across diverse applications. Machine unlearning, which focuses on efficiently removing specific data influences to comply with privacy regulations, shows promise in restricting updates to influence-critical parameters. However, existing parameter-efficient unlearning methods are largely devised in a module-oblivious manner, which tends to inaccurately identify these parameters and leads to inferior unlearning performance for Transformers. In this paper, we propose {\tt MAPE-Unlearn}, a module-aware parameter-efficient machine unlearning approach that uses a learnable pair of masks to pinpoint influence-critical parameters in the heads and filters of Transformers. The learning objective of these masks is derived by desiderata of unlearning and optimized through an efficient algorithm featured by a greedy search with a warm start. Extensive experiments on various Transformer models and datasets demonstrate the effectiveness and robustness of {\tt MAPE-Unlearn} for unlearning.
Similar Papers
Improving Unlearning with Model Updates Probably Aligned with Gradients
Machine Learning (CS)
Removes specific data from AI without breaking it.
UIPE: Enhancing LLM Unlearning by Removing Knowledge Related to Forgetting Targets
Computation and Language
Cleans harmful knowledge from AI without breaking it.
RapidUn: Influence-Driven Parameter Reweighting for Efficient Large Language Model Unlearning
Computation and Language
Teaches AI to forget bad information quickly.