Evaluating Multilingual and Code-Switched Alignment in LLMs via Synthetic Natural Language Inference
By: Samir Abdaljalil , Erchin Serpedin , Khalid Qaraqe and more
Potential Business Impact:
Makes computers understand different languages better.
Large language models (LLMs) are increasingly applied in multilingual contexts, yet their capacity for consistent, logically grounded alignment across languages remains underexplored. We present a controlled evaluation framework for multilingual natural language inference (NLI) that generates synthetic, logic-based premise-hypothesis pairs and translates them into a typologically diverse set of languages. This design enables precise control over semantic relations and allows testing in both monolingual and mixed-language (code-switched) conditions. Surprisingly, code-switching does not degrade, and can even improve, performance, suggesting that translation-induced lexical variation may serve as a regularization signal. We validate semantic preservation through embedding-based similarity analyses and cross-lingual alignment visualizations, confirming the fidelity of translated pairs. Our findings expose both the potential and the brittleness of current LLM cross-lingual reasoning, and identify code-switching as a promising lever for improving multilingual robustness. Code available at: https://github.com/KurbanIntelligenceLab/nli-stress-testing
Similar Papers
Breaking Language Barriers: Equitable Performance in Multilingual Language Models
Computation and Language
Helps computers understand less common languages better.
Beyond Monolingual Assumptions: A Survey of Code-Switched NLP in the Era of Large Language Models
Computation and Language
Helps computers understand mixed-language conversations.
Investigating and Scaling up Code-Switching for Multilingual Language Model Pre-Training
Computation and Language
Helps computers learn many languages by mixing them.