Erasing CLIP Memories: Non-Destructive, Data-Free Zero-Shot class Unlearning in CLIP Models
By: Ashish Mishra , Tarun Kumar , Gyanaranjan Nayak and more
Potential Business Impact:
Removes unwanted knowledge from AI models.
We introduce a novel, closed-form approach for selective unlearning in multimodal models, specifically targeting pretrained models such as CLIP. Our method leverages nullspace projection to erase the target class information embedded in the final projection layer, without requiring any retraining or the use of images from the forget set. By computing an orthonormal basis for the subspace spanned by target text embeddings and projecting these directions, we dramatically reduce the alignment between image features and undesired classes. Unlike traditional unlearning techniques that rely on iterative fine-tuning and extensive data curation, our approach is both computationally efficient and surgically precise. This leads to a pronounced drop in zero-shot performance for the target classes while preserving the overall multimodal knowledge of the model. Our experiments demonstrate that even a partial projection can balance between complete unlearning and retaining useful information, addressing key challenges in model decontamination and privacy preservation.
Similar Papers
Selective, Controlled and Domain-Agnostic Unlearning in Pretrained CLIP: A Training- and Data-Free Approach
CV and Pattern Recognition
Removes unwanted knowledge from AI without retraining.
Targeted Forgetting of Image Subgroups in CLIP Models
CV and Pattern Recognition
Cleans AI's bad memories without hurting good ones.
Memory-Free Continual Learning with Null Space Adaptation for Zero-Shot Vision-Language Models
Artificial Intelligence
Helps AI learn new things without forgetting old ones.