Explanation Beyond Intuition: A Testable Criterion for Inherent Explainability
By: Michael Merry, Pat Riddle, Jim Warren
Inherent explainability is the gold standard in Explainable Artificial Intelligence (XAI). However, there is not a consistent definition or test to demonstrate inherent explainability. Work to date either characterises explainability through metrics, or appeals to intuition - "we know it when we see it". We propose a globally applicable criterion for inherent explainability. The criterion uses graph theory for representing and decomposing models for structure-local explanation, and recomposing them into global explanations. We form the structure-local explanations as annotations, a verifiable hypothesis-evidence structure that allows for a range of explanatory methods to be used. This criterion matches existing intuitions on inherent explainability, and provides justifications why a large regression model may not be explainable but a sparse neural network could be. We differentiate explainable -- a model that allows for explanation -- and \textit{explained} -- one that has a verified explanation. Finally, we provide a full explanation of PREDICT -- a Cox proportional hazards model of cardiovascular disease risk, which is in active clinical use in New Zealand. It follows that PREDICT is inherently explainable. This work provides structure to formalise other work on explainability, and allows regulators a flexible but rigorous test that can be used in compliance frameworks.
Similar Papers
A Framework for Causal Concept-based Model Explanations
Artificial Intelligence
Explains how AI makes decisions using simple ideas.
How can we trust opaque systems? Criteria for robust explanations in XAI
Machine Learning (CS)
Makes smart computer guesses understandable and trustworthy.
Explainable artificial intelligence (XAI): from inherent explainability to large language models
Machine Learning (CS)
Lets people understand why computers make choices.