Toward Scalable and Valid Conditional Independence Testing with Spectral Representations
By: Alek Frohlich , Vladimir Kostic , Karim Lounici and more
Conditional independence (CI) is central to causal inference, feature selection, and graphical modeling, yet it is untestable in many settings without additional assumptions. Existing CI tests often rely on restrictive structural conditions, limiting their validity on real-world data. Kernel methods using the partial covariance operator offer a more principled approach but suffer from limited adaptivity, slow convergence, and poor scalability. In this work, we explore whether representation learning can help address these limitations. Specifically, we focus on representations derived from the singular value decomposition of the partial covariance operator and use them to construct a simple test statistic, reminiscent of the Hilbert-Schmidt Independence Criterion (HSIC). We also introduce a practical bi-level contrastive algorithm to learn these representations. Our theory links representation learning error to test performance and establishes asymptotic validity and power guarantees. Preliminary experiments suggest that this approach offers a practical and statistically grounded path toward scalable CI testing, bridging kernel-based theory with modern representation learning.
Similar Papers
On the Hardness of Conditional Independence Testing In Practice
Machine Learning (Stat)
Finds why computer tests for fairness sometimes fail.
Testing Conditional Independence via the Spectral Generalized Covariance Measure: Beyond Euclidean Data
Methodology
Finds hidden connections between data points.
Testing Conditional Independence via the Spectral Generalized Covariance Measure: Beyond Euclidean Data
Methodology
Finds hidden connections between data points.