Linguistic Neuron Overlap Patterns to Facilitate Cross-lingual Transfer on Low-resource Languages
By: Yuemei Xu , Kexin Xu , Jian Zhou and more
Potential Business Impact:
Helps computers understand rare languages better.
The current Large Language Models (LLMs) face significant challenges in improving performance on low-resource languages and urgently need data-efficient methods without costly fine-tuning. From the perspective of language-bridge, we propose BridgeX-ICL, a simple yet effective method to improve zero-shot Cross-lingual In-Context Learning (X-ICL) for low-resource languages. Unlike existing works focusing on language-specific neurons, BridgeX-ICL explores whether sharing neurons can improve cross-lingual performance in LLMs or not. We construct neuron probe data from the ground-truth MUSE bilingual dictionaries, and define a subset of language overlap neurons accordingly, to ensure full activation of these anchored neurons. Subsequently, we propose an HSIC-based metric to quantify LLMs' internal linguistic spectrum based on overlap neurons, which guides optimal bridge selection. The experiments conducted on 2 cross-lingual tasks and 15 language pairs from 7 diverse families (covering both high-low and moderate-low pairs) validate the effectiveness of BridgeX-ICL and offer empirical insights into the underlying multilingual mechanisms of LLMs.
Similar Papers
Code-Switching In-Context Learning for Cross-Lingual Transfer of Large Language Models
Computation and Language
Helps computers understand many languages better.
Cross-Lingual Generalization and Compression: From Language-Specific to Shared Neurons
Computation and Language
Computers learn to understand words in many languages.
Enhancing LLM Language Adaption through Cross-lingual In-Context Pre-training
Computation and Language
Helps computers understand many languages better.