Fine-Grained Bias Exploration and Mitigation for Group-Robust Classification
By: Miaoyun Zhao, Qiang Zhang, Chenrong Li
Potential Business Impact:
Helps computers learn without unfair biases.
Achieving group-robust generalization in the presence of spurious correlations remains a significant challenge, particularly when bias annotations are unavailable. Recent studies on Class-Conditional Distribution Balancing (CCDB) reveal that spurious correlations often stem from mismatches between the class-conditional and marginal distributions of bias attributes. They achieve promising results by addressing this issue through simple distribution matching in a bias-agnostic manner. However, CCDB approximates each distribution using a single Gaussian, which is overly simplistic and rarely holds in real-world applications. To address this limitation, we propose a novel method called Bias Exploration via Overfitting (BEO), which captures each distribution in greater detail by modeling it as a mixture of latent groups. Building on these group-level descriptions, we introduce a fine-grained variant of CCDB, termed FG-CCDB, which performs more precise distribution matching and balancing within each group. Through group-level reweighting, FG-CCDB learns sample weights from a global perspective, achieving stronger mitigation of spurious correlations without incurring substantial storage or computational costs. Extensive experiments demonstrate that BEO serves as a strong proxy for ground-truth bias annotations and can be seamlessly integrated with bias-supervised methods. Moreover, when combined with FG-CCDB, our method performs on par with bias-supervised approaches on binary classification tasks and significantly outperforms them in highly biased multi-class scenarios.
Similar Papers
Class-Conditional Distribution Balancing for Group Robust Classification
Machine Learning (CS)
Fixes computer guesses that are wrong for bad reasons.
Improving Group Robustness on Spurious Correlation via Evidential Alignment
Machine Learning (CS)
Teaches computers to see things correctly, not by guessing.
Mitigating Clever Hans Strategies in Image Classifiers through Generating Counterexamples
Machine Learning (CS)
Teaches computers to learn better, not just guess.