Assessing Intersectional Bias in Representations of Pre-Trained Image Recognition Models
By: Valerie Krug, Sebastian Stober
Potential Business Impact:
Finds unfairness in computer face recognition.
Deep Learning models have achieved remarkable success. Training them is often accelerated by building on top of pre-trained models which poses the risk of perpetuating encoded biases. Here, we investigate biases in the representations of commonly used ImageNet classifiers for facial images while considering intersections of sensitive variables age, race and gender. To assess the biases, we use linear classifier probes and visualize activations as topographic maps. We find that representations in ImageNet classifiers particularly allow differentiation between ages. Less strongly pronounced, the models appear to associate certain ethnicities and distinguish genders in middle-aged groups.
Similar Papers
Biased Heritage: How Datasets Shape Models in Facial Expression Recognition
CV and Pattern Recognition
Makes AI understand faces without unfairness.
Hidden Bias in the Machine: Stereotypes in Text-to-Image Models
CV and Pattern Recognition
Shows how AI pictures can show unfair ideas.
T2IBias: Uncovering Societal Bias Encoded in the Latent Space of Text-to-Image Generative Models
Machine Learning (CS)
AI makes pictures show unfair stereotypes.