CoUn: Empowering Machine Unlearning via Contrastive Learning
By: Yasser H. Khalil, Mehdi Setayesh, Hongliang Li
Potential Business Impact:
Removes bad data from computer brains.
Machine unlearning (MU) aims to remove the influence of specific "forget" data from a trained model while preserving its knowledge of the remaining "retain" data. Existing MU methods based on label manipulation or model weight perturbations often achieve limited unlearning effectiveness. To address this, we introduce CoUn, a novel MU framework inspired by the observation that a model retrained from scratch using only retain data classifies forget data based on their semantic similarity to the retain data. CoUn emulates this behavior by adjusting learned data representations through contrastive learning (CL) and supervised learning, applied exclusively to retain data. Specifically, CoUn (1) leverages semantic similarity between data samples to indirectly adjust forget representations using CL, and (2) maintains retain representations within their respective clusters through supervised learning. Extensive experiments across various datasets and model architectures show that CoUn consistently outperforms state-of-the-art MU baselines in unlearning effectiveness. Additionally, integrating our CL module into existing baselines empowers their unlearning effectiveness.
Similar Papers
Group-robust Machine Unlearning
Machine Learning (CS)
Makes AI forget some data fairly.
Deep Contrastive Unlearning for Language Models
Computation and Language
Removes private info from AI without breaking it.
Concept Unlearning in Large Language Models via Self-Constructed Knowledge Triplets
Computation and Language
Removes specific ideas from AI, not just words.