Exploring the Feasibility of Multilingual Grammatical Error Correction with a Single LLM up to 9B parameters: A Comparative Study of 17 Models
By: Dawid Wisniewski, Antoni Solarski, Artur Nowakowski
Potential Business Impact:
Fixes grammar in many languages with one tool.
Recent language models can successfully solve various language-related tasks, and many understand inputs stated in different languages. In this paper, we explore the performance of 17 popular models used to correct grammatical issues in texts stated in English, German, Italian, and Swedish when using a single model to correct texts in all those languages. We analyze the outputs generated by these models, focusing on decreasing the number of grammatical errors while keeping the changes small. The conclusions drawn help us understand what problems occur among those models and which models can be recommended for multilingual grammatical error correction tasks. We list six models that improve grammatical correctness in all four languages and show that Gemma 9B is currently the best performing one for the languages considered.
Similar Papers
Multilingual Machine Translation with Open Large Language Models at Practical Scale: An Empirical Study
Computation and Language
Makes computers translate 28 languages perfectly.
Multilingual Performance Biases of Large Language Models in Education
Computation and Language
Tests if computers help students learn other languages.
Exploring Robustness of Multilingual LLMs on Real-World Noisy Data
Computation and Language
Computers learn to understand misspelled words.