Machine-Learning-Assisted Comparison of Regression Functions
By: Jian Yan , Zhuoxi Li , Yang Ning and more
Potential Business Impact:
Compares data patterns even with many details.
We revisit the classical problem of comparing regression functions, a fundamental question in statistical inference with broad relevance to modern applications such as data integration, transfer learning, and causal inference. Existing approaches typically rely on smoothing techniques and are thus hindered by the curse of dimensionality. We propose a generalized notion of kernel-based conditional mean dependence that provides a new characterization of the null hypothesis of equal regression functions. Building on this reformulation, we develop two novel tests that leverage modern machine learning methods for flexible estimation. We establish the asymptotic properties of the test statistics, which hold under both fixed- and high-dimensional regimes. Unlike existing methods that often require restrictive distributional assumptions, our framework only imposes mild moment conditions. The efficacy of the proposed tests is demonstrated through extensive numerical studies.
Similar Papers
A kernel conditional two-sample test
Machine Learning (CS)
Finds when two groups of data are different.
Practically significant differences between conditional distribution functions
Econometrics
Tests if two groups are different enough.
Kernel-Based Nonparametric Tests For Shape Constraints
Machine Learning (Stat)
Helps make better money choices with math.