Sequential Subspace Noise Injection Prevents Accuracy Collapse in Certified Unlearning
By: Polina Dolgova, Sebastian U. Stich
Potential Business Impact:
Removes data from AI without hurting its smarts.
Certified unlearning based on differential privacy offers strong guarantees but remains largely impractical: the noisy fine-tuning approaches proposed so far achieve these guarantees but severely reduce model accuracy. We propose sequential noise scheduling, which distributes the noise budget across orthogonal subspaces of the parameter space, rather than injecting it all at once. This simple modification mitigates the destructive effect of noise while preserving the original certification guarantees. We extend the analysis of noisy fine-tuning to the subspace setting, proving that the same $(\varepsilon,δ)$ privacy budget is retained. Empirical results on image classification benchmarks show that our approach substantially improves accuracy after unlearning while remaining robust to membership inference attacks. These results show that certified unlearning can achieve both rigorous guarantees and practical utility.
Similar Papers
Gaussian Certified Unlearning in High Dimensions: A Hypothesis Testing Approach
Machine Learning (Stat)
Removes unwanted data from AI without hurting its smarts.
Parameter-Efficient Fine-Tuning with Differential Privacy for Robust Instruction Adaptation in Large Language Models
Computation and Language
Keeps AI learning private and fast.
Inducing Uncertainty for Test-Time Privacy
Machine Learning (CS)
Makes AI forget data, even when it tries.