Robust Fake News Detection using Large Language Models under Adversarial Sentiment Attacks
By: Sahar Tahmasebi, Eric Müller-Budack, Ralph Ewerth
Potential Business Impact:
Stops fake news from tricking computers by changing feelings.
Misinformation and fake news have become a pressing societal challenge, driving the need for reliable automated detection methods. Prior research has highlighted sentiment as an important signal in fake news detection, either by analyzing which sentiments are associated with fake news or by using sentiment and emotion features for classification. However, this poses a vulnerability since adversaries can manipulate sentiment to evade detectors especially with the advent of large language models (LLMs). A few studies have explored adversarial samples generated by LLMs, but they mainly focus on stylistic features such as writing style of news publishers. Thus, the crucial vulnerability of sentiment manipulation remains largely unexplored. In this paper, we investigate the robustness of state-of-the-art fake news detectors under sentiment manipulation. We introduce AdSent, a sentiment-robust detection framework designed to ensure consistent veracity predictions across both original and sentiment-altered news articles. Specifically, we (1) propose controlled sentiment-based adversarial attacks using LLMs, (2) analyze the impact of sentiment shifts on detection performance. We show that changing the sentiment heavily impacts the performance of fake news detection models, indicating biases towards neutral articles being real, while non-neutral articles are often classified as fake content. (3) We introduce a novel sentiment-agnostic training strategy that enhances robustness against such perturbations. Extensive experiments on three benchmark datasets demonstrate that AdSent significantly outperforms competitive baselines in both accuracy and robustness, while also generalizing effectively to unseen datasets and adversarial scenarios.
Similar Papers
Model-Agnostic Sentiment Distribution Stability Analysis for Robust LLM-Generated Texts Detection
Computation and Language
Finds fake writing by checking emotions.
FactGuard: Event-Centric and Commonsense-Guided Fake News Detection
Artificial Intelligence
Finds fake news by checking facts, not just writing.
Adversarial News and Lost Profits: Manipulating Headlines in LLM-Driven Algorithmic Trading
Cryptography and Security
Tricks computers into losing money with fake news.