Interpretable Model-Aware Counterfactual Explanations for Random Forest
By: Joshua S. Harvey , Guanchao Feng , Sai Anusha Meesala and more
Potential Business Impact:
Explains why computer decisions change outcomes.
Despite their enormous predictive power, machine learning models are often unsuitable for applications in regulated industries such as finance, due to their limited capacity to provide explanations. While model-agnostic frameworks such as Shapley values have proved to be convenient and popular, they rarely align with the kinds of causal explanations that are typically sought after. Counterfactual case-based explanations, where an individual is informed of which circumstances would need to be different to cause a change in outcome, may be more intuitive and actionable. However, finding appropriate counterfactual cases is an open challenge, as is interpreting which features are most critical for the change in outcome. Here, we pose the question of counterfactual search and interpretation in terms of similarity learning, exploiting the representation learned by the random forest predictive model itself. Once a counterfactual is found, the feature importance of the explanation is computed as a function of which random forest partitions are crossed in order to reach it from the original instance. We demonstrate this method on both the MNIST hand-drawn digit dataset and the German credit dataset, finding that it generates explanations that are sparser and more useful than Shapley values.
Similar Papers
Graph Diffusion Counterfactual Explanation
Machine Learning (CS)
Helps AI explain why it makes graph decisions.
Counterfactual Explanations for Model Ensembles Using Entropic Risk Measures
Machine Learning (CS)
Helps AI decisions work for many AI models.
Comparative Explanations via Counterfactual Reasoning in Recommendations
Information Retrieval
Shows why you like something by comparing it.