Fake News Classification in Urdu: A Domain Adaptation Approach for a Low-Resource Language
By: Muhammad Zain Ali, Bernhard Pfahringer, Tony Smith
Potential Business Impact:
Helps find fake news in Urdu.
Misinformation on social media is a widely acknowledged issue, and researchers worldwide are actively engaged in its detection. However, low-resource languages such as Urdu have received limited attention in this domain. An obvious approach is to utilize a multilingual pretrained language model and fine-tune it for a downstream classification task, such as misinformation detection. However, these models struggle with domain-specific terms, leading to suboptimal performance. To address this, we investigate the effectiveness of domain adaptation before fine-tuning for fake news classification in Urdu, employing a staged training approach to optimize model generalization. We evaluate two widely used multilingual models, XLM-RoBERTa and mBERT, and apply domain-adaptive pretraining using a publicly available Urdu news corpus. Experiments on four publicly available Urdu fake news datasets show that domain-adapted XLM-R consistently outperforms its vanilla counterpart, while domain-adapted mBERT exhibits mixed results.
Similar Papers
Unified Large Language Models for Misinformation Detection in Low-Resource Linguistic Settings
Computation and Language
Helps find fake news in Urdu.
AI-Generated Text Detection in Low-Resource Languages: A Case Study on Urdu
Computation and Language
Finds fake writing in Urdu.
From Scarcity to Capability: Empowering Fake News Detection in Low-Resource Languages with LLMs
Computation and Language
Helps stop fake news in Bangla language.