Dimension-Free Correlated Sampling for the Hypersimplex
By: Joseph , Naor , Nitya Raju and more
Potential Business Impact:
Makes computer programs share information better.
Sampling from multiple distributions so as to maximize overlap has been studied by statisticians since the 1950s. Since the 2000s, such correlated sampling from the probability simplex has been a powerful building block in disparate areas of theoretical computer science. We study a generalization of this problem to sampling sets from given vectors in the hypersimplex, i.e., outputting sets of size (at most) some $k$ in $[n]$, while maximizing the sampled sets' overlap. Specifically, the expected difference between two output sets should be at most $α$ times their input vectors' $\ell_1$ distance. A value of $α=O(\log n)$ is known to be achievable, due to Chen et al.~(ICALP'17). We improve this factor to $O(\log k)$, independent of the ambient dimension~$n$. Our algorithm satisfies other desirable properties, including (up to a $\log^* n$ factor) input-sparsity sampling time, logarithmic parallel depth and dynamic update time, as well as preservation of submodular objectives. Anticipating broader use of correlated sampling algorithms for the hypersimplex, we present applications of our algorithm to online paging, offline approximation of metric multi-labeling and swift multi-scenario submodular welfare approximating reallocation.
Similar Papers
Improved Sample Complexity for Full Coverage in Compact and Continuous Spaces
Machine Learning (CS)
Finds more things with fewer tries.
Hypothesis Selection: A High Probability Conundrum
Data Structures and Algorithms
Finds the best data explanation faster.
A simple analysis of a quantum-inspired algorithm for solving low-rank linear systems
Data Structures and Algorithms
Finds answers to math problems much faster.