Backdoor Attacks on Multi-modal Contrastive Learning
By: Simi D Kuniyilh, Rita Machacy
Potential Business Impact:
Protects AI from secret sabotage during training.
Contrastive learning has become a leading self- supervised approach to representation learning across domains, including vision, multimodal settings, graphs, and federated learning. However, recent studies have shown that contrastive learning is susceptible to backdoor and data poisoning attacks. In these attacks, adversaries can manipulate pretraining data or model updates to insert hidden malicious behavior. This paper offers a thorough and comparative review of backdoor attacks in contrastive learning. It analyzes threat models, attack methods, target domains, and available defenses. We summarize recent advancements in this area, underline the specific vulnerabilities inherent to contrastive learning, and discuss the challenges and future research directions. Our findings have significant implications for the secure deployment of systems in industrial and distributed environments.
Similar Papers
Robust Defense Strategies for Multimodal Contrastive Learning: Efficient Fine-tuning Against Backdoor Attacks
CV and Pattern Recognition
Finds and fixes hidden "bad code" in AI.
Backdoor Attacks and Defenses in Computer Vision Domain: A Survey
Cryptography and Security
Protects smart cameras from secret tricks.
BackdoorVLM: A Benchmark for Backdoor Attacks on Vision-Language Models
CV and Pattern Recognition
Finds hidden tricks in AI that can fool it.