Data-Efficient Prediction-Powered Calibration via Cross-Validation
By: Seonghoon Yoo , Houssem Sifaou , Sangwoo Park and more
Potential Business Impact:
Makes AI decisions more trustworthy with less data.
Calibration data are necessary to formally quantify the uncertainty of the decisions produced by an existing artificial intelligence (AI) model. To overcome the common issue of scarce calibration data, a promising approach is to employ synthetic labels produced by a (generally different) predictive model. However, fine-tuning the label-generating predictor on the inference task of interest, as well as estimating the residual bias of the synthetic labels, demand additional data, potentially exacerbating the calibration data scarcity problem. This paper introduces a novel approach that efficiently utilizes limited calibration data to simultaneously fine-tune a predictor and estimate the bias of the synthetic labels. The proposed method yields prediction sets with rigorous coverage guarantees for AI-generated decisions. Experimental results on an indoor localization problem validate the effectiveness and performance gains of our solution.
Similar Papers
Calibration improves detection of mislabeled examples
Machine Learning (CS)
Fixes computer learning mistakes in messy data.
Monitoring the calibration of probability forecasts with an application to concept drift detection involving image classification
Machine Learning (Stat)
Keeps computer vision accurate over time.
When Models Know When They Do Not Know: Calibration, Cascading, and Cleaning
Artificial Intelligence
Helps AI know when it's wrong.