Rigorous Feature Importance Scores based on Shapley Value and Banzhaf Index
By: Xuanxiang Huang, Olivier Létoffé, Joao Marques-Silva
Potential Business Impact:
Helps AI understand why it's wrong.
Feature attribution methods based on game theory are ubiquitous in the field of eXplainable Artificial Intelligence (XAI). Recent works proposed rigorous feature attribution using logic-based explanations, specifically targeting high-stakes uses of machine learning (ML) models. Typically, such works exploit weak abductive explanation (WAXp) as the characteristic function to assign importance to features. However, one possible downside is that the contribution of non-WAXp sets is neglected. In fact, non-WAXp sets can also convey important information, because of the relationship between formal explanations (XPs) and adversarial examples (AExs). Accordingly, this paper leverages Shapley value and Banzhaf index to devise two novel feature importance scores. We take into account non-WAXp sets when computing feature contribution, and the novel scores quantify how effective each feature is at excluding AExs. Furthermore, the paper identifies properties and studies the computational complexity of the proposed scores.
Similar Papers
Explaining Risks: Axiomatic Risk Attributions for Financial Models
Computational Finance
Explains why computer guesses are right or wrong.
Attribution Score Alignment in Explainable Data Management
Databases
Helps find the most important data for answers.
Evaluation of Black-Box XAI Approaches for Predictors of Values of Boolean Formulae
Artificial Intelligence
Helps AI explain its decisions fairly.