Score: 1

Forgetting Similar Samples: Can Machine Unlearning Do it Better?

Published: January 11, 2026 | arXiv ID: 2601.06938v1

By: Heng Xu , Tianqing Zhu , Dayong Ye and more

Potential Business Impact:

Removes unwanted data's effect from AI.

Business Areas:
Machine Learning Artificial Intelligence, Data and Analytics, Software

Machine unlearning, a process enabling pre-trained models to remove the influence of specific training samples, has attracted significant attention in recent years. Although extensive research has focused on developing efficient machine unlearning strategies, we argue that these methods mainly aim at removing samples rather than removing samples' influence on the model, thus overlooking the fundamental definition of machine unlearning. In this paper, we first conduct a comprehensive study to evaluate the effectiveness of existing unlearning schemes when the training dataset includes many samples similar to those targeted for unlearning. Specifically, we evaluate: Do existing unlearning methods truly adhere to the original definition of machine unlearning and effectively eliminate all influence of target samples when similar samples are present in the training dataset? Our extensive experiments, conducted on four carefully constructed datasets with thorough analysis, reveal a notable gap between the expected and actual performance of most existing unlearning methods for image and language models, even for the retraining-from-scratch baseline. Additionally, we also explore potential solutions to enhance current unlearning approaches.

Country of Origin
🇨🇳 China


Page Count
28 pages

Category
Computer Science:
Machine Learning (CS)