Forgetting Similar Samples: Can Machine Unlearning Do it Better?
By: Heng Xu , Tianqing Zhu , Dayong Ye and more
Potential Business Impact:
Removes unwanted data's effect from AI.
Machine unlearning, a process enabling pre-trained models to remove the influence of specific training samples, has attracted significant attention in recent years. Although extensive research has focused on developing efficient machine unlearning strategies, we argue that these methods mainly aim at removing samples rather than removing samples' influence on the model, thus overlooking the fundamental definition of machine unlearning. In this paper, we first conduct a comprehensive study to evaluate the effectiveness of existing unlearning schemes when the training dataset includes many samples similar to those targeted for unlearning. Specifically, we evaluate: Do existing unlearning methods truly adhere to the original definition of machine unlearning and effectively eliminate all influence of target samples when similar samples are present in the training dataset? Our extensive experiments, conducted on four carefully constructed datasets with thorough analysis, reveal a notable gap between the expected and actual performance of most existing unlearning methods for image and language models, even for the retraining-from-scratch baseline. Additionally, we also explore potential solutions to enhance current unlearning approaches.
Similar Papers
When unlearning is free: leveraging low influence points to reduce computational costs
Machine Learning (CS)
Cleans computer brains faster by ignoring tiny details.
Toward Reliable Machine Unlearning: Theory, Algorithms, and Evaluation
Machine Learning (CS)
Makes AI forget specific information safely.
Not All Data Are Unlearned Equally
Computation and Language
Removes unwanted information from AI minds.