Toward a Safer Web: Multilingual Multi-Agent LLMs for Mitigating Adversarial Misinformation Attacks
By: Nouar Aldahoul, Yasir Zaki
Potential Business Impact:
Fights fake news by spotting tricky language tricks.
The rapid spread of misinformation on digital platforms threatens public discourse, emotional stability, and decision-making. While prior work has explored various adversarial attacks in misinformation detection, the specific transformations examined in this paper have not been systematically studied. In particular, we investigate language-switching across English, French, Spanish, Arabic, Hindi, and Chinese, followed by translation. We also study query length inflation preceding summarization and structural reformatting into multiple-choice questions. In this paper, we present a multilingual, multi-agent large language model framework with retrieval-augmented generation that can be deployed as a web plugin into online platforms. Our work underscores the importance of AI-driven misinformation detection in safeguarding online factual integrity against diverse attacks, while showcasing the feasibility of plugin-based deployment for real-world web applications.
Similar Papers
Toward Verifiable Misinformation Detection: A Multi-Tool LLM Agent Framework
Artificial Intelligence
Finds fake news by checking facts online.
A Multilingual, Large-Scale Study of the Interplay between LLM Safeguards, Personalisation, and Disinformation
Computation and Language
AI can create fake news tailored to you.
A Survey on Proactive Defense Strategies Against Misinformation in Large Language Models
Information Retrieval
Stops AI from spreading fake news.