Score: 1

Beyond Weaponization: NLP Security for Medium and Lower-Resourced Languages in Their Own Right

Published: July 4, 2025 | arXiv ID: 2507.03473v1

By: Heather Lent

Potential Business Impact:

Makes AI safer for all languages, not just English.

Business Areas:
Natural Language Processing Artificial Intelligence, Data and Analytics, Software

Despite mounting evidence that multilinguality can be easily weaponized against language models (LMs), works across NLP Security remain overwhelmingly English-centric. In terms of securing LMs, the NLP norm of "English first" collides with standard procedure in cybersecurity, whereby practitioners are expected to anticipate and prepare for worst-case outcomes. To mitigate worst-case outcomes in NLP Security, researchers must be willing to engage with the weakest links in LM security: lower-resourced languages. Accordingly, this work examines the security of LMs for lower- and medium-resourced languages. We extend existing adversarial attacks for up to 70 languages to evaluate the security of monolingual and multilingual LMs for these languages. Through our analysis, we find that monolingual models are often too small in total number of parameters to ensure sound security, and that while multilinguality is helpful, it does not always guarantee improved security either. Ultimately, these findings highlight important considerations for more secure deployment of LMs, for communities of lower-resourced languages.

Country of Origin
🇩🇰 Denmark

Repos / Data Links

Page Count
19 pages

Category
Computer Science:
Computation and Language