CaliMatch: Adaptive Calibration for Improving Safe Semi-supervised Learning
By: Jinsoo Bae, Seoung Bum Kim, Hyungrok Do
Potential Business Impact:
Makes AI better at learning with less labeled data.
Semi-supervised learning (SSL) uses unlabeled data to improve the performance of machine learning models when labeled data is scarce. However, its real-world applications often face the label distribution mismatch problem, in which the unlabeled dataset includes instances whose ground-truth labels are absent from the labeled training dataset. Recent studies, referred to as safe SSL, have addressed this issue by using both classification and out-of-distribution (OOD) detection. However, the existing methods may suffer from overconfidence in deep neural networks, leading to increased SSL errors because of high confidence in incorrect pseudo-labels or OOD detection. To address this, we propose a novel method, CaliMatch, which calibrates both the classifier and the OOD detector to foster safe SSL. CaliMatch presents adaptive label smoothing and temperature scaling, which eliminates the need to manually tune the smoothing degree for effective calibration. We give a theoretical justification for why improving the calibration of both the classifier and the OOD detector is crucial in safe SSL. Extensive evaluations on CIFAR-10, CIFAR-100, SVHN, TinyImageNet, and ImageNet demonstrate that CaliMatch outperforms the existing methods in safe SSL tasks.
Similar Papers
CalibrateMix: Guided-Mixup Calibration of Image Semi-Supervised Models
CV and Pattern Recognition
Makes computer guesses more honest and accurate.
Sampling Control for Imbalanced Calibration in Semi-Supervised Learning
Machine Learning (CS)
Fixes computer learning when some groups are rare.
CGMatch: A Different Perspective of Semi-supervised Learning
CV and Pattern Recognition
Teaches computers better with less examples.