Score: 0

Evaluating Cross-Lingual Unlearning in Multilingual Language Models

Published: January 10, 2026 | arXiv ID: 2601.06675v1

By: Tyler Lizzo, Larry Heck

Potential Business Impact:

Removes unwanted info from AI in all languages.

Business Areas:
Language Learning Education

We present the first comprehensive evaluation of cross-lingual unlearning in multilingual LLMs. Using translated TOFU benchmarks in seven language/script variants, we test major unlearning algorithms and show that most fail to remove facts outside the training language, even when utility remains high. However, subspace-projection consistently outperforms the other methods, achieving strong cross-lingual forgetting with minimal degradation. Analysis of learned task subspaces reveals a shared interlingua structure: removing this shared subspace harms all languages, while removing language-specific components selectively affects one. These results demonstrate that multilingual forgetting depends on geometry in weight space, motivating subspace-based approaches for future unlearning systems.

Country of Origin
πŸ‡ΊπŸ‡Έ United States

Page Count
18 pages

Category
Computer Science:
Computation and Language