A Tale of Two Classes: Adapting Supervised Contrastive Learning to Binary Imbalanced Datasets
By: David Mildenberger , Paul Hager , Daniel Rueckert and more
Potential Business Impact:
Improves AI's ability to spot rare diseases.
Supervised contrastive learning (SupCon) has proven to be a powerful alternative to the standard cross-entropy loss for classification of multi-class balanced datasets. However, it struggles to learn well-conditioned representations of datasets with long-tailed class distributions. This problem is potentially exacerbated for binary imbalanced distributions, which are commonly encountered during many real-world problems such as medical diagnosis. In experiments on seven binary datasets of natural and medical images, we show that the performance of SupCon decreases with increasing class imbalance. To substantiate these findings, we introduce two novel metrics that evaluate the quality of the learned representation space. By measuring the class distribution in local neighborhoods, we are able to uncover structural deficiencies of the representation space that classical metrics cannot detect. Informed by these insights, we propose two new supervised contrastive learning strategies tailored to binary imbalanced datasets that improve the structure of the representation space and increase downstream classification accuracy over standard SupCon by up to 35%. We make our code available.
Similar Papers
Variational Supervised Contrastive Learning
Machine Learning (CS)
Makes computer images more organized and understandable.
Generalizing Supervised Contrastive learning: A Projection Perspective
Machine Learning (CS)
Teaches computers to learn from examples better.
WeakSupCon: Weakly Supervised Contrastive Learning for Encoder Pre-training
CV and Pattern Recognition
Teaches computers to learn from pictures with hidden labels.