Poisoned-MRAG: Knowledge Poisoning Attacks to Multimodal Retrieval Augmented Generation
By: Yinuo Liu , Zenghui Yuan , Guiyao Tie and more
Potential Business Impact:
Makes AI show wrong answers by tricking its memory.
Multimodal retrieval-augmented generation (RAG) enhances the visual reasoning capability of vision-language models (VLMs) by dynamically accessing information from external knowledge bases. In this work, we introduce \textit{Poisoned-MRAG}, the first knowledge poisoning attack on multimodal RAG systems. Poisoned-MRAG injects a few carefully crafted image-text pairs into the multimodal knowledge database, manipulating VLMs to generate the attacker-desired response to a target query. Specifically, we formalize the attack as an optimization problem and propose two cross-modal attack strategies, dirty-label and clean-label, tailored to the attacker's knowledge and goals. Our extensive experiments across multiple knowledge databases and VLMs show that Poisoned-MRAG outperforms existing methods, achieving up to 98\% attack success rate with just five malicious image-text pairs injected into the InfoSeek database (481,782 pairs). Additionally, We evaluate 4 different defense strategies, including paraphrasing, duplicate removal, structure-driven mitigation, and purification, demonstrating their limited effectiveness and trade-offs against Poisoned-MRAG. Our results highlight the effectiveness and scalability of Poisoned-MRAG, underscoring its potential as a significant threat to multimodal RAG systems.
Similar Papers
One Pic is All it Takes: Poisoning Visual Document Retrieval Augmented Generation with a Single Image
Computation and Language
Makes AI lie by tricking its memory.
Practical Poisoning Attacks against Retrieval-Augmented Generation
Cryptography and Security
Makes AI smarter and harder to trick.
Defending Against Knowledge Poisoning Attacks During Retrieval-Augmented Generation
Machine Learning (CS)
Stops bad info from tricking smart computer programs.