Score: 1

Improving Multilingual Math Reasoning for African Languages

Published: May 26, 2025 | arXiv ID: 2505.19848v1

By: Odunayo Ogundepo , Akintunde Oladipo , Kelechi Ogueji and more

Potential Business Impact:

Helps computers understand African languages better.

Business Areas:
Natural Language Processing Artificial Intelligence, Data and Analytics, Software

Researchers working on low-resource languages face persistent challenges due to limited data availability and restricted access to computational resources. Although most large language models (LLMs) are predominantly trained in high-resource languages, adapting them to low-resource contexts, particularly African languages, requires specialized techniques. Several strategies have emerged for adapting models to low-resource languages in todays LLM landscape, defined by multi-stage pre-training and post-training paradigms. However, the most effective approaches remain uncertain. This work systematically investigates which adaptation strategies yield the best performance when extending existing LLMs to African languages. We conduct extensive experiments and ablation studies to evaluate different combinations of data types (translated versus synthetically generated), training stages (pre-training versus post-training), and other model adaptation configurations. Our experiments focuses on mathematical reasoning tasks, using the Llama 3.1 model family as our base model.


Page Count
18 pages

Category
Computer Science:
Computation and Language