Score: 2

Enhancing Arabic Automated Essay Scoring with Synthetic Data and Error Injection

Published: March 22, 2025 | arXiv ID: 2503.17739v2

By: Chatrine Qwaider , Bashar Alhafni , Kirill Chirkunov and more

Potential Business Impact:

Teaches computers to grade Arabic essays better.

Business Areas:
Natural Language Processing Artificial Intelligence, Data and Analytics, Software

Automated Essay Scoring (AES) plays a crucial role in assessing language learners' writing quality, reducing grading workload, and providing real-time feedback. The lack of annotated essay datasets inhibits the development of Arabic AES systems. This paper leverages Large Language Models (LLMs) and Transformer models to generate synthetic Arabic essays for AES. We prompt an LLM to generate essays across the Common European Framework of Reference (CEFR) proficiency levels and introduce and compare two approaches to error injection. We create a dataset of 3,040 annotated essays with errors injected using our two methods. Additionally, we develop a BERT-based Arabic AES system calibrated to CEFR levels. Our experimental results demonstrate the effectiveness of our synthetic dataset in improving Arabic AES performance. We make our code and data publicly available.

Country of Origin
🇦🇪 🇺🇸 United States, United Arab Emirates

Repos / Data Links

Page Count
15 pages

Category
Computer Science:
Computation and Language