A Practical Introduction to Kernel Discrepancies: MMD, HSIC & KSD
By: Antonin Schrab
Potential Business Impact:
Measures how different two groups of data are.
This article provides a practical introduction to kernel discrepancies, focusing on the Maximum Mean Discrepancy (MMD), the Hilbert-Schmidt Independence Criterion (HSIC), and the Kernel Stein Discrepancy (KSD). Various estimators for these discrepancies are presented, including the commonly-used V-statistics and U-statistics, as well as several forms of the more computationally-efficient incomplete U-statistics. The importance of the choice of kernel bandwidth is stressed, showing how it affects the behaviour of the discrepancy estimation. Adaptive estimators are introduced, which combine multiple estimators with various kernels, addressing the problem of kernel selection.
Similar Papers
Integral-Operator-Based Spectral Algorithms for Goodness-of-Fit Tests
Methodology
Makes computer tests better at spotting fake data.
A Martingale Kernel Two-Sample Test
Methodology
Finds differences between groups faster.
Optimizing Kernel Discrepancies via Subset Selection
Machine Learning (Stat)
Finds better computer samples for math problems.