Score: 1

Pretraining Finnish ModernBERTs

Published: November 12, 2025 | arXiv ID: 2511.09213v1

By: Akseli Reunamo , Laura-Maria Peltonen , Hans Moen and more

Potential Business Impact:

Makes computers understand Finnish text better.

Business Areas:
Natural Language Processing Artificial Intelligence, Data and Analytics, Software

This paper reports on pretraining ModernBERT encoder models in six different sizes, ranging from 51M to 475M parameters, with a focus on limited multilingualism, emphasizing languages relevant to Finland. Our models are competitive with, or superior to, existing multilingual models. They outperform monolingual models on tasks that require a context longer than 512 tokens. We present empirical results on using different data in the final stage of training. The code and models are publicly released.

Country of Origin
🇫🇮 Finland


Page Count
21 pages

Category
Computer Science:
Computation and Language