Evidence-Grounded Multimodal Misinformation Detection with Attention-Based GNNs
By: Sharad Duwal , Mir Nafis Sharear Shopnil , Abhishek Tyagi and more
Potential Business Impact:
Finds fake news by checking pictures and words.
Multimodal out-of-context (OOC) misinformation is misinformation that repurposes real images with unrelated or misleading captions. Detecting such misinformation is challenging because it requires resolving the context of the claim before checking for misinformation. Many current methods, including LLMs and LVLMs, do not perform this contextualization step. LLMs hallucinate in absence of context or parametric knowledge. In this work, we propose a graph-based method that evaluates the consistency between the image and the caption by constructing two graph representations: an evidence graph, derived from online textual evidence, and a claim graph, from the claim in the caption. Using graph neural networks (GNNs) to encode and compare these representations, our framework then evaluates the truthfulness of image-caption pairs. We create datasets for our graph-based method, evaluate and compare our baseline model against popular LLMs on the misinformation detection task. Our method scores $93.05\%$ detection accuracy on the evaluation set and outperforms the second-best performing method (an LLM) by $2.82\%$, making a case for smaller and task-specific methods.
Similar Papers
EXCLAIM: An Explainable Cross-Modal Agentic System for Misinformation Detection with Hierarchical Retrieval
Information Retrieval
Finds fake news by checking pictures and words.
Mitigating GenAI-powered Evidence Pollution for Out-of-Context Multimodal Misinformation Detection
Multimedia
Find fake news even when AI tries to hide it.
E2LVLM:Evidence-Enhanced Large Vision-Language Model for Multimodal Out-of-Context Misinformation Detection
Machine Learning (CS)
Stops fake news by checking images and text.