Score: 2

BabyBabelLM: A Multilingual Benchmark of Developmentally Plausible Training Data

Published: October 11, 2025 | arXiv ID: 2510.10159v1

By: Jaap Jumelet , Abdellah Fourtassi , Akari Haga and more

Potential Business Impact:

Teaches computers to learn languages like babies.

Business Areas:
Natural Language Processing Artificial Intelligence, Data and Analytics, Software

We present BabyBabelLM, a multilingual collection of datasets modeling the language a person observes from birth until they acquire a native language. We curate developmentally plausible pretraining data aiming to cover the equivalent of 100M English words of content in each of 45 languages. We compile evaluation suites and train baseline models in each language. BabyBabelLM aims to facilitate multilingual pretraining and cognitive modeling.

Country of Origin
🇮🇱 🇳🇱 Israel, Netherlands

Repos / Data Links

Page Count
33 pages

Category
Computer Science:
Computation and Language