Score: 0

Backdoor Attacks on Multi-modal Contrastive Learning

Published: January 16, 2026 | arXiv ID: 2601.11006v1

By: Simi D Kuniyilh, Rita Machacy

Potential Business Impact:

Protects AI from secret sabotage during training.

Business Areas:
Machine Learning Artificial Intelligence, Data and Analytics, Software

Contrastive learning has become a leading self- supervised approach to representation learning across domains, including vision, multimodal settings, graphs, and federated learning. However, recent studies have shown that contrastive learning is susceptible to backdoor and data poisoning attacks. In these attacks, adversaries can manipulate pretraining data or model updates to insert hidden malicious behavior. This paper offers a thorough and comparative review of backdoor attacks in contrastive learning. It analyzes threat models, attack methods, target domains, and available defenses. We summarize recent advancements in this area, underline the specific vulnerabilities inherent to contrastive learning, and discuss the challenges and future research directions. Our findings have significant implications for the secure deployment of systems in industrial and distributed environments.

Page Count
8 pages

Category
Computer Science:
Machine Learning (CS)