Improving Unlearning with Model Updates Probably Aligned with Gradients
By: Virgile Dine, Teddy Furon, Charly Faure
Potential Business Impact:
Removes specific data from AI without breaking it.
We formulate the machine unlearning problem as a general constrained optimization problem. It unifies the first-order methods from the approximate machine unlearning literature. This paper then introduces the concept of feasible updates as the model's parameter update directions that help with unlearning while not degrading the utility of the initial model. Our design of feasible updates is based on masking, \ie\ a careful selection of the model's parameters worth updating. It also takes into account the estimation noise of the gradients when processing each batch of data to offer a statistical guarantee to derive locally feasible updates. The technique can be plugged in, as an add-on, to any first-order approximate unlearning methods. Experiments with computer vision classifiers validate this approach.
Similar Papers
Module-Aware Parameter-Efficient Machine Unlearning on Transformers
Machine Learning (CS)
Removes unwanted data from AI without breaking it.
Toward Reliable Machine Unlearning: Theory, Algorithms, and Evaluation
Machine Learning (CS)
Makes AI forget specific information safely.
Provable Unlearning with Gradient Ascent on Two-Layer ReLU Neural Networks
Machine Learning (CS)
Removes private data from AI without retraining.