Score: 1

Improving Multilingual Language Models by Aligning Representations through Steering

Published: May 19, 2025 | arXiv ID: 2505.12584v2

By: Omar Mahmoud , Buddhika Laknath Semage , Thommen George Karimpanal and more

Potential Business Impact:

Makes computers understand many languages better.

Business Areas:
Natural Language Processing Artificial Intelligence, Data and Analytics, Software

This paper investigates how Large Language Models (LLMs) represent non-English tokens -- a question that remains underexplored despite recent progress. We propose a lightweight intervention method using representation steering, where a learned vector is added to the residual stream at a single model layer to enhance multilingual performance. Through extensive experiments across seven competitive baselines -- including prompt optimization, supervised fine-tuning (SFT), in-context learning, cross-lingual transfer, and translation-based methods-we show that our approach consistently outperforms most alternatives. In particular, it achieves performance on par with production-grade translation systems while requiring far fewer resources. We further explore the complementarity between our method and SFT, demonstrating that steering offers a direct, efficient way to realign internal representations. These findings underscore the potential of activation-level interventions as a powerful tool for improving the multilingual capabilities of LLMs.

Country of Origin
🇦🇺 Australia

Repos / Data Links

Page Count
21 pages

Category
Computer Science:
Computation and Language