Composite goodness-of-fit test with the Kernel Stein Discrepancy and a bootstrap for degenerate U-statistics with estimated parameters
By: Florian Brueck, Veronika Reimoser, Fabian Baier
Potential Business Impact:
Checks if computer models fit real-world data.
This paper formally derives the asymptotic distribution of a goodness-of-fit test based on the Kernel Stein Discrepancy introduced in (Oscar Key et al., "Composite Goodness-of-fit Tests with Kernels", Journal of Machine Learning Research 26.51 (2025), pp. 1-60). The test enables the simultaneous estimation of the optimal parameter within a parametric family of candidate models. Its asymptotic distribution is shown to be a weighted sum of infinitely many $\chi^2$-distributed random variables plus an additional disturbance term, which is due to the parameter estimation. Further, we provide a general framework to bootstrap degenerate parameter-dependent $U$-statistics and use it to derive a new Kernel Stein Discrepancy composite goodness-of-fit test.
Similar Papers
Optimal-Transport Based Multivariate Goodness-of-Fit Tests
Statistics Theory
Checks if data matches a pattern.
Semiparametric KSD test: unifying score and distance-based approaches for goodness-of-fit testing
Machine Learning (Stat)
Tests if computer models are good enough.
Fast and Scalable Score-Based Kernel Calibration Tests
Machine Learning (Stat)
Checks if computer predictions are trustworthy.