Survey of Adversarial Robustness in Multimodal Large Language Models
By: Chengze Jiang , Zhuangzhuang Wang , Minjing Dong and more
Potential Business Impact:
Makes AI understand pictures and words safely.
Multimodal Large Language Models (MLLMs) have demonstrated exceptional performance in artificial intelligence by facilitating integrated understanding across diverse modalities, including text, images, video, audio, and speech. However, their deployment in real-world applications raises significant concerns about adversarial vulnerabilities that could compromise their safety and reliability. Unlike unimodal models, MLLMs face unique challenges due to the interdependencies among modalities, making them susceptible to modality-specific threats and cross-modal adversarial manipulations. This paper reviews the adversarial robustness of MLLMs, covering different modalities. We begin with an overview of MLLMs and a taxonomy of adversarial attacks tailored to each modality. Next, we review key datasets and evaluation metrics used to assess the robustness of MLLMs. After that, we provide an in-depth review of attacks targeting MLLMs across different modalities. Our survey also identifies critical challenges and suggests promising future research directions.
Similar Papers
Investigating Vulnerabilities and Defenses Against Audio-Visual Attacks: A Comprehensive Survey Emphasizing Multimodal Models
Cryptography and Security
Makes AI that sees and hears unsafe.
Adversarial Attacks in Multimodal Systems: A Practitioner's Survey
Machine Learning (CS)
Protects smart AI from being tricked.
A Survey of Attacks on Large Language Models
Cryptography and Security
Protects smart computer programs from being tricked.