Optimizing Kernel Discrepancies via Subset Selection
By: Deyao Chen , François Clément , Carola Doerr and more
Potential Business Impact:
Finds better computer samples for math problems.
Kernel discrepancies are a powerful tool for analyzing worst-case errors in quasi-Monte Carlo (QMC) methods. Building on recent advances in optimizing such discrepancy measures, we extend the subset selection problem to the setting of kernel discrepancies, selecting an m-element subset from a large population of size $n \gg m$. We introduce a novel subset selection algorithm applicable to general kernel discrepancies to efficiently generate low-discrepancy samples from both the uniform distribution on the unit hypercube, the traditional setting of classical QMC, and from more general distributions $F$ with known density functions by employing the kernel Stein discrepancy. We also explore the relationship between the classical $L_2$ star discrepancy and its $L_\infty$ counterpart.
Similar Papers
A Practical Introduction to Kernel Discrepancies: MMD, HSIC & KSD
Machine Learning (Stat)
Measures how different two groups of data are.
High-Dimensional Quasi-Monte Carlo via Combinatorial Discrepancy
Numerical Analysis
Makes computer math faster for hard problems.
Fast and Scalable Score-Based Kernel Calibration Tests
Machine Learning (Stat)
Checks if computer predictions are trustworthy.