A Robust Certified Machine Unlearning Method Under Distribution Shift
By: Jinduo Guo, Yinzhi Cao
Potential Business Impact:
Makes AI forget data even when it's not random.
The Newton method has been widely adopted to achieve certified unlearning. A critical assumption in existing approaches is that the data requested for unlearning are selected i.i.d.(independent and identically distributed). However,the problem of certified unlearning under non-i.i.d. deletions remains largely unexplored. In practice, unlearning requests are inherently biased, leading to non-i.i.d. deletions and causing distribution shifts between the original and retained datasets. In this paper, we show that certified unlearning with the Newton method becomes inefficient and ineffective under non-i.i.d. unlearning sets. We then propose a better certified unlearning approach by performing a distribution-aware certified unlearning framework based on iterative Newton updates constrained by a trust region. Our method provides a closer approximation to the retrained model and yields a tighter pre-run bound on the gradient residual, thereby ensuring efficient (epsilon, delta)-certified unlearning. To demonstrate its practical effectiveness under distribution shift, we also conduct extensive experiments across multiple evaluation metrics, providing a comprehensive assessment of our approach.
Similar Papers
Certified Unlearning in Decentralized Federated Learning
Machine Learning (CS)
Removes your data from shared AI models.
Fully Decentralized Certified Unlearning
Machine Learning (CS)
Removes private data from AI without retraining.
Certified Data Removal Under High-dimensional Settings
Machine Learning (Stat)
Removes unwanted info from AI models.