Conditional cross-fitting for unbiased machine-learning-assisted covariate adjustment in randomized experiments
By: Xin Lu , Lei Shi , Hanzhong Liu and more
Potential Business Impact:
Makes study results more accurate with less data.
Randomized experiments are the gold standard for estimating the average treatment effect (ATE). While covariate adjustment can reduce the asymptotic variances of the unbiased Horvitz-Thompson estimators for the ATE, it suffers from finite-sample biases due to data reuse in both prediction and estimation. Traditional sample-splitting and cross-fitting methods can address the problem of data reuse and obtain unbiased estimators. However, they require that the data are independently and identically distributed, which is usually violated under the design-based inference framework for randomized experiments. To address this challenge, we propose a novel conditional cross-fitting method, under the design-based inference framework, where potential outcomes and covariates are fixed and the randomization is the sole source of randomness. We propose sample-splitting algorithms for various randomized experiments, including Bernoulli randomized experiments, completely randomized experiments, and stratified randomized experiments. Based on the proposed algorithms, we construct unbiased covariate-adjusted ATE estimators and propose valid inference procedures. Our methods can accommodate flexible machine-learning-assisted covariate adjustments and allow for model misspecification.
Similar Papers
Design-based finite-sample analysis for regression adjustment
Statistics Theory
Makes study results more accurate, even with lots of data.
Regression adjustment in covariate-adaptive randomized experiments with missing covariates
Methodology
Fixes missing data in medical tests for better results.
Cross-Balancing for Data-Informed Design and Efficient Analysis of Observational Studies
Methodology
Makes studies fairer by using results to pick groups.