A Unified View of Optimal Kernel Hypothesis Testing
By: Antonin Schrab
Potential Business Impact:
Finds patterns in data to make smart guesses.
This paper provides a unifying view of optimal kernel hypothesis testing across the MMD two-sample, HSIC independence, and KSD goodness-of-fit frameworks. Minimax optimal separation rates in the kernel and $L^2$ metrics are presented, with two adaptive kernel selection methods (kernel pooling and aggregation), and under various testing constraints: computational efficiency, differential privacy, and robustness to data corruption. Intuition behind the derivation of the power results is provided in a unified way accross the three frameworks, and open problems are highlighted.
Similar Papers
DUAL: Learning Diverse Kernels for Aggregated Two-sample and Independence Testing
Machine Learning (CS)
Finds better patterns in messy data.
A kernel conditional two-sample test
Machine Learning (CS)
Finds when two groups of data are different.
Nonparametric hazard rate estimation with associated kernels and minimax bandwidth choice
Statistics Theory
Predicts how long things will last.