Score: 0

$φ$-test: Global Feature Selection and Inference for Shapley Additive Explanations

Published: December 8, 2025 | arXiv ID: 2512.07578v1

By: Dongseok Kim , Hyoungsun Choi , Mohamed Jismy Aashik Rasool and more

Potential Business Impact:

Finds the most important parts of computer predictions.

Business Areas:
A/B Testing Data and Analytics

We propose $φ$-test, a global feature-selection and significance procedure for black-box predictors that combines Shapley attributions with selective inference. Given a trained model and an evaluation dataset, $φ$-test performs SHAP-guided screening and fits a linear surrogate on the screened features via a selection rule with a tractable selective-inference form. For each retained feature, it outputs a Shapley-based global score, a surrogate coefficient, and post-selection $p$-values and confidence intervals in a global feature-importance table. Experiments on real tabular regression tasks with tree-based and neural backbones suggest that $φ$-test can retain much of the predictive ability of the original model while using only a few features and producing feature sets that remain fairly stable across resamples and backbone classes. In these settings, $φ$-test acts as a practical global explanation layer linking Shapley-based importance summaries with classical statistical inference.

Country of Origin
🇰🇷 Korea, Republic of

Page Count
15 pages

Category
Statistics:
Machine Learning (Stat)