Score: 1

Conformal Risk Training: End-to-End Optimization of Conformal Risk Control

Published: October 9, 2025 | arXiv ID: 2510.08748v1

By: Christopher Yeh , Nicolas Christianson , Adam Wierman and more

Potential Business Impact:

Guarantees AI predictions are safe and reliable.

Business Areas:
Risk Management Professional Services

While deep learning models often achieve high predictive accuracy, their predictions typically do not come with any provable guarantees on risk or reliability, which are critical for deployment in high-stakes applications. The framework of conformal risk control (CRC) provides a distribution-free, finite-sample method for controlling the expected value of any bounded monotone loss function and can be conveniently applied post-hoc to any pre-trained deep learning model. However, many real-world applications are sensitive to tail risks, as opposed to just expected loss. In this work, we develop a method for controlling the general class of Optimized Certainty-Equivalent (OCE) risks, a broad class of risk measures which includes as special cases the expected loss (generalizing the original CRC method) and common tail risks like the conditional value-at-risk (CVaR). Furthermore, standard post-hoc CRC can degrade average-case performance due to its lack of feedback to the model. To address this, we introduce "conformal risk training," an end-to-end approach that differentiates through conformal OCE risk control during model training or fine-tuning. Our method achieves provable risk guarantees while demonstrating significantly improved average-case performance over post-hoc approaches on applications to controlling classifiers' false negative rate and controlling financial risk in battery storage operation.

Country of Origin
🇺🇸 United States

Repos / Data Links

Page Count
30 pages

Category
Computer Science:
Machine Learning (CS)