Ridge Boosting is Both Robust and Efficient
By: David Bruns-Smith, Zhongming Xie, Avi Feller
Potential Business Impact:
Makes one computer model work well for many jobs.
Estimators in statistics and machine learning must typically trade off between efficiency, having low variance for a fixed target, and distributional robustness, such as \textit{multiaccuracy}, or having low bias over a range of possible targets. In this paper, we consider a simple estimator, \emph{ridge boosting}: starting with any initial predictor, perform a single boosting step with (kernel) ridge regression. Surprisingly, we show that ridge boosting simultaneously achieves both efficiency and distributional robustness: for target distribution shifts that lie within an RKHS unit ball, this estimator maintains low bias across all such shifts and has variance at the semiparametric efficiency bound for each target. In addition to bridging otherwise distinct research areas, this result has immediate practical value. Since ridge boosting uses only data from the source distribution, researchers can train a single model to obtain both robust and efficient estimates for multiple target estimands at the same time, eliminating the need to fit separate semiparametric efficient estimators for each target. We assess this approach through simulations and an application estimating the age profile of retirement income.
Similar Papers
An RKHS Perspective on Tree Ensembles
Machine Learning (Stat)
Explains why computer learning models work so well.
Transfer Learning Across Fixed-Income Product Classes
Machine Learning (Stat)
Helps predict money values more accurately.
Statistical Inference for Gradient Boosting Regression
Machine Learning (Stat)
Makes computer predictions more trustworthy and accurate.