Multimedia Verification Through Multi-Agent Deep Research Multimodal Large Language Models
By: Huy Hoan Le , Van Sy Thinh Nguyen , Thi Le Chi Dang and more
Potential Business Impact:
Finds fake videos and pictures online.
This paper presents our submission to the ACMMM25 - Grand Challenge on Multimedia Verification. We developed a multi-agent verification system that combines Multimodal Large Language Models (MLLMs) with specialized verification tools to detect multimedia misinformation. Our system operates through six stages: raw data processing, planning, information extraction, deep research, evidence collection, and report generation. The core Deep Researcher Agent employs four tools: reverse image search, metadata analysis, fact-checking databases, and verified news processing that extracts spatial, temporal, attribution, and motivational context. We demonstrate our approach on a challenge dataset sample involving complex multimedia content. Our system successfully verified content authenticity, extracted precise geolocation and timing information, and traced source attribution across multiple platforms, effectively addressing real-world multimedia verification scenarios.
Similar Papers
DeepMMSearch-R1: Empowering Multimodal LLMs in Multimodal Web Search
CV and Pattern Recognition
Lets computers search the web for answers.
Fact-Checking at Scale: Multimodal AI for Authenticity and Context Verification in Online Media
Multimedia
Checks if online videos and pictures are real.
Fact-Checking at Scale: Multimodal AI for Authenticity and Context Verification in Online Media
Multimedia
Checks if online videos and pictures are real.