Multimodal Zero-Shot Framework for Deepfake Hate Speech Detection in Low-Resource Languages
By: Rishabh Ranjan , Likhith Ayinala , Mayank Vatsa and more
Potential Business Impact:
Finds hate speech in fake voices, even new ones.
This paper introduces a novel multimodal framework for hate speech detection in deepfake audio, excelling even in zero-shot scenarios. Unlike previous approaches, our method uses contrastive learning to jointly align audio and text representations across languages. We present the first benchmark dataset with 127,290 paired text and synthesized speech samples in six languages: English and five low-resource Indian languages (Hindi, Bengali, Marathi, Tamil, Telugu). Our model learns a shared semantic embedding space, enabling robust cross-lingual and cross-modal classification. Experiments on two multilingual test sets show our approach outperforms baselines, achieving accuracies of 0.819 and 0.701, and generalizes well to unseen languages. This demonstrates the advantage of combining modalities for hate speech detection in synthetic media, especially in low-resource settings where unimodal models falter. The Dataset is available at https://www.iab-rubric.org/resources.
Similar Papers
SynHate: Detecting Hate Speech in Synthetic Deepfake Audio
Sound
Finds fake hate speech in any language.
Few-shot Hate Speech Detection Based on the MindSpore Framework
Computation and Language
Finds hate speech with less examples.
cantnlp@DravidianLangTech2025: A Bag-of-Sounds Approach to Multimodal Hate Speech Detection
Computation and Language
Detects hateful speech in online videos and audio.