Improving the statistical efficiency of cross-conformal prediction
By: Matteo Gasparin, Aaditya Ramdas
Potential Business Impact:
Makes computer guesses more accurate and smaller.
Vovk (2015) introduced cross-conformal prediction, a modification of split conformal designed to improve the width of prediction sets. The method, when trained with a miscoverage rate equal to $\alpha$ and $n \gg K$, ensures a marginal coverage of at least $1 - 2\alpha - 2(1-\alpha)(K-1)/(n+K)$, where $n$ is the number of observations and $K$ denotes the number of folds. A simple modification of the method achieves coverage of at least $1-2\alpha$. In this work, we propose new variants of both methods that yield smaller prediction sets without compromising the latter theoretical guarantees. The proposed methods are based on recent results deriving more statistically efficient combination of p-values that leverage exchangeability and randomization. Simulations confirm the theoretical findings and bring out some important tradeoffs.
Similar Papers
A Few Observations on Sample-Conditional Coverage in Conformal Prediction
Statistics Theory
Makes predictions more trustworthy and accurate.
Conditional validity and a fast approximation formula of full conformal prediction sets
Statistics Theory
Makes predictions more reliable with less math.
Probabilistic Conformal Coverage Guarantees in Small-Data Settings
Machine Learning (CS)
Guarantees predictions are correct more often.