Empirical Likelihood Meets Prediction-Powered Inference
By: Guanghui Wang, Mengtao Wen, Changliang Zou
We study inference with a small labeled sample, a large unlabeled sample, and high-quality predictions from an external model. We link prediction-powered inference with empirical likelihood by stacking supervised estimating equations based on labeled outcomes with auxiliary moment conditions built from predictions, and then optimizing empirical likelihood under these joint constraints. The resulting empirical likelihood-based prediction-powered inference (EPI) estimator is asymptotically normal, has asymptotic variance no larger than the fully supervised estimator, and attains the semiparametric efficiency bound when the auxiliary functions span the predictable component of the supervised score. For hypothesis testing and confidence sets, empirical likelihood ratio statistics admit chi-squared-type limiting distributions. As a by-product, the empirical likelihood weights induce a calibrated empirical distribution that integrates supervised and prediction-based information, enabling estimation and uncertainty quantification for general functionals beyond parameters defined by estimating equations. We present two practical implementations: one based on basis expansions in the predictions and covariates, and one that learns an approximately optimal auxiliary function by cross-fitting. In simulations and applications, EPI reduces mean squared error and shortens confidence intervals while maintaining nominal coverage.
Similar Papers
Expectation-propagation for Bayesian empirical likelihood inference
Methodology
Makes computer guesses more accurate without needing exact rules.
Prediction-Powered Inference with Inverse Probability Weighting
Machine Learning (Stat)
Improves guessing with less data.
Pseudo Empirical Likelihood Inference for Non-Probability Survey Samples
Methodology
Improves how we learn from surveys with missing info.