BUSTED at AraGenEval Shared Task: A Comparative Study of Transformer-Based Models for Arabic AI-Generated Text Detection
By: Ali Zain, Sareem Farooqui, Muhammad Rafi
Potential Business Impact:
Finds fake Arabic writing using smart computer programs.
This paper details our submission to the Ara- GenEval Shared Task on Arabic AI-generated text detection, where our team, BUSTED, se- cured 5th place. We investigated the effec- tiveness of three pre-trained transformer mod- els: AraELECTRA, CAMeLBERT, and XLM- RoBERTa. Our approach involved fine-tuning each model on the provided dataset for a binary classification task. Our findings revealed a sur- prising result: the multilingual XLM-RoBERTa model achieved the highest performance with an F1 score of 0.7701, outperforming the spe- cialized Arabic models. This work underscores the complexities of AI-generated text detection and highlights the strong generalization capa- bilities of multilingual models.
Similar Papers
!MSA at BAREC Shared Task 2025: Ensembling Arabic Transformers for Readability Assessment
Computation and Language
Helps computers understand hard Arabic text better.
AI-Generated Text Detection in Low-Resource Languages: A Case Study on Urdu
Computation and Language
Finds fake writing in Urdu.
PolyTruth: Multilingual Disinformation Detection using Transformer-Based Language Models
Computation and Language
AI spots fake news in many languages.