Bias Analysis in Unconditional Image Generative Models
By: Xiaofeng Zhang , Michelle Lin , Simon Lacoste-Julien and more
Potential Business Impact:
Finds how AI unfairly shows some things.
The widespread adoption of generative AI models has raised growing concerns about representational harm and potential discriminatory outcomes. Yet, despite growing literature on this topic, the mechanisms by which bias emerges - especially in unconditional generation - remain disentangled. We define the bias of an attribute as the difference between the probability of its presence in the observed distribution and its expected proportion in an ideal reference distribution. In our analysis, we train a set of unconditional image generative models and adopt a commonly used bias evaluation framework to study bias shift between training and generated distributions. Our experiments reveal that the detected attribute shifts are small. We find that the attribute shifts are sensitive to the attribute classifier used to label generated images in the evaluation framework, particularly when its decision boundaries fall in high-density regions. Our empirical analysis indicates that this classifier sensitivity is often observed in attributes values that lie on a spectrum, as opposed to exhibiting a binary nature. This highlights the need for more representative labeling practices, understanding the shortcomings through greater scrutiny of evaluation frameworks, and recognizing the socially complex nature of attributes when evaluating bias.
Similar Papers
Exploring Bias in over 100 Text-to-Image Generative Models
CV and Pattern Recognition
Finds how AI art tools become unfair over time.
Manifold Induced Biases for Zero-shot and Few-shot Detection of Generated Images
CV and Pattern Recognition
Finds fake pictures made by computers.
AI Alignment in Medical Imaging: Unveiling Hidden Biases Through Counterfactual Analysis
Machine Learning (CS)
Checks if AI doctors are fair to everyone.