Learning Intersections of Two Margin Halfspaces under Factorizable Distributions
By: Ilias Diakonikolas , Mingchen Ma , Lisheng Ren and more
Potential Business Impact:
Teaches computers to learn patterns faster.
Learning intersections of halfspaces is a central problem in Computational Learning Theory. Even for just two halfspaces, it remains a major open question whether learning is possible in polynomial time with respect to the margin $γ$ of the data points and their dimensionality $d$. The best-known algorithms run in quasi-polynomial time $d^{O(\log(1/γ))}$, and it has been shown that this complexity is unavoidable for any algorithm relying solely on correlational statistical queries (CSQ). In this work, we introduce a novel algorithm that provably circumvents the CSQ hardness barrier. Our approach applies to a broad class of distributions satisfying a natural, previously studied, factorizability assumption. Factorizable distributions lie between distribution-specific and distribution-free settings, and significantly extend previously known tractable cases. Under these distributions, we show that CSQ-based methods still require quasipolynomial time even for weakly learning, whereas our algorithm achieves $poly(d,1/γ)$ time by leveraging more general statistical queries (SQ), establishing a strong separation between CSQ and SQ for this simple realizable PAC learning problem. Our result is grounded in a rigorous analysis utilizing a novel duality framework that characterizes the moment tensor structure induced by the marginal distributions. Building on these structural insights, we propose new, efficient learning algorithms. These algorithms combine a refined variant of Jennrich's Algorithm with PCA over random projections of the moment tensor, along with a gradient-descent-based non-convex optimization framework.
Similar Papers
Learning Intersections of Halfspaces under Factorizable Distribution
Machine Learning (CS)
Finds patterns faster than old methods.
A Fully Polynomial-Time Algorithm for Robustly Learning Halfspaces over the Hypercube
Data Structures and Algorithms
Teaches computers to learn from messy, bad data.
Smoothed Agnostic Learning of Halfspaces over the Hypercube
Machine Learning (CS)
Learns patterns in computer code faster.