Score: 1

Stability Regularized Cross-Validation

Published: May 11, 2025 | arXiv ID: 2505.06927v1

By: Ryan Cory-Wright, Andrés Gómez

Potential Business Impact:

Makes computer models more reliable on new data.

Business Areas:
A/B Testing Data and Analytics

We revisit the problem of ensuring strong test-set performance via cross-validation. Motivated by the generalization theory literature, we propose a nested k-fold cross-validation scheme that selects hyperparameters by minimizing a weighted sum of the usual cross-validation metric and an empirical model-stability measure. The weight on the stability term is itself chosen via a nested cross-validation procedure. This reduces the risk of strong validation set performance and poor test set performance due to instability. We benchmark our procedure on a suite of 13 real-world UCI datasets, and find that, compared to k-fold cross-validation over the same hyperparameters, it improves the out-of-sample MSE for sparse ridge regression and CART by 4% on average, but has no impact on XGBoost. This suggests that for interpretable and unstable models, such as sparse regression and CART, our approach is a viable and computationally affordable method for improving test-set performance.

Country of Origin
🇺🇸 🇬🇧 United States, United Kingdom

Page Count
23 pages

Category
Mathematics:
Optimization and Control