Harnessing Diffusion-Generated Synthetic Images for Fair Image Classification
By: Abhipsa Basu , Aviral Gupta , Abhijnya Bhat and more
Potential Business Impact:
Makes AI fairer by fixing biased training pictures.
Image classification systems often inherit biases from uneven group representation in training data. For example, in face datasets for hair color classification, blond hair may be disproportionately associated with females, reinforcing stereotypes. A recent approach leverages the Stable Diffusion model to generate balanced training data, but these models often struggle to preserve the original data distribution. In this work, we explore multiple diffusion-finetuning techniques, e.g., LoRA and DreamBooth, to generate images that more accurately represent each training group by learning directly from their samples. Additionally, in order to prevent a single DreamBooth model from being overwhelmed by excessive intra-group variations, we explore a technique of clustering images within each group and train a DreamBooth model per cluster. These models are then used to generate group-balanced data for pretraining, followed by fine-tuning on real data. Experiments on multiple benchmarks demonstrate that the studied finetuning approaches outperform vanilla Stable Diffusion on average and achieve results comparable to SOTA debiasing techniques like Group-DRO, while surpassing them as the dataset bias severity increases.
Similar Papers
Diffusing DeBias: Synthetic Bias Amplification for Model Debiasing
Machine Learning (CS)
Fixes computer "thinking" errors caused by bad examples.
Debiasing Diffusion Model: Enhancing Fairness through Latent Representation Learning in Stable Diffusion Model
Machine Learning (CS)
Makes AI art fair for everyone.
Investigation of Accuracy and Bias in Face Recognition Trained with Synthetic Data
CV and Pattern Recognition
Makes face recognition fairer with fake pictures.