$φ$-test: Global Feature Selection and Inference for Shapley Additive Explanations
By: Dongseok Kim , Hyoungsun Choi , Mohamed Jismy Aashik Rasool and more
Potential Business Impact:
Finds the most important parts of computer predictions.
We propose $φ$-test, a global feature-selection and significance procedure for black-box predictors that combines Shapley attributions with selective inference. Given a trained model and an evaluation dataset, $φ$-test performs SHAP-guided screening and fits a linear surrogate on the screened features via a selection rule with a tractable selective-inference form. For each retained feature, it outputs a Shapley-based global score, a surrogate coefficient, and post-selection $p$-values and confidence intervals in a global feature-importance table. Experiments on real tabular regression tasks with tree-based and neural backbones suggest that $φ$-test can retain much of the predictive ability of the original model while using only a few features and producing feature sets that remain fairly stable across resamples and backbone classes. In these settings, $φ$-test acts as a practical global explanation layer linking Shapley-based importance summaries with classical statistical inference.
Similar Papers
ScoresActivation: A New Activation Function for Model Agnostic Global Explainability by Design
Machine Learning (CS)
Makes AI explain its choices while learning.
SHAP Distance: An Explainability-Aware Metric for Evaluating the Semantic Fidelity of Synthetic Tabular Data
Machine Learning (CS)
Checks if fake data teaches computers right lessons.
Rigorous Feature Importance Scores based on Shapley Value and Banzhaf Index
Artificial Intelligence
Helps AI understand why it's wrong.