CALM: Culturally Self-Aware Language Models
By: Lingzhi Shen , Xiaohao Cai , Yunfei Long and more
Potential Business Impact:
Teaches computers to understand different cultures.
Cultural awareness in language models is the capacity to understand and adapt to diverse cultural contexts. However, most existing approaches treat culture as static background knowledge, overlooking its dynamic and evolving nature. This limitation reduces their reliability in downstream tasks that demand genuine cultural sensitivity. In this work, we introduce CALM, a novel framework designed to endow language models with cultural self-awareness. CALM disentangles task semantics from explicit cultural concepts and latent cultural signals, shaping them into structured cultural clusters through contrastive learning. These clusters are then aligned via cross-attention to establish fine-grained interactions among related cultural features and are adaptively integrated through a Mixture-of-Experts mechanism along culture-specific dimensions. The resulting unified representation is fused with the model's original knowledge to construct a culturally grounded internal identity state, which is further enhanced through self-prompted reflective learning, enabling continual adaptation and self-correction. Extensive experiments conducted on multiple cross-cultural benchmark datasets demonstrate that CALM consistently outperforms state-of-the-art methods.
Similar Papers
CALM: Unleashing the Cross-Lingual Self-Aligning Ability of Language Model Question Answering
Computation and Language
Makes computers answer same questions in any language.
Cultural Learning-Based Culture Adaptation of Language Models
Computation and Language
Teaches computers to understand different cultures.
Meta-Cultural Competence: Climbing the Right Hill of Cultural Awareness
Computers and Society
AI learns to understand any culture, not just one.