Cultural Awareness in Vision-Language Models: A Cross-Country Exploration
By: Avinash Madasu, Vasudev Lal, Phillip Howard
Potential Business Impact:
Finds how computers see people and places unfairly.
Vision-Language Models (VLMs) are increasingly deployed in diverse cultural contexts, yet their internal biases remain poorly understood. In this work, we propose a novel framework to systematically evaluate how VLMs encode cultural differences and biases related to race, gender, and physical traits across countries. We introduce three retrieval-based tasks: (1) Race to Country retrieval, which examines the association between individuals from specific racial groups (East Asian, White, Middle Eastern, Latino, South Asian, and Black) and different countries; (2) Personal Traits to Country retrieval, where images are paired with trait-based prompts (e.g., Smart, Honest, Criminal, Violent) to investigate potential stereotypical associations; and (3) Physical Characteristics to Country retrieval, focusing on visual attributes like skinny, young, obese, and old to explore how physical appearances are culturally linked to nations. Our findings reveal persistent biases in VLMs, highlighting how visual representations may inadvertently reinforce societal stereotypes.
Similar Papers
Uncovering Cultural Representation Disparities in Vision-Language Models
CV and Pattern Recognition
Finds AI's unfair views on different countries.
CultureVLM: Characterizing and Improving Cultural Understanding of Vision-Language Models for over 100 Countries
Artificial Intelligence
Teaches AI to understand cultures worldwide.
Visual Cues of Gender and Race are Associated with Stereotyping in Vision-Language Models
CV and Pattern Recognition
Makes AI tell more uniform stories about women.