A Framework for Causal Concept-based Model Explanations
By: Anna Rodum Bjøru , Jacob Lysnæs-Larsen , Oskar Jørgensen and more
Potential Business Impact:
Explains how AI makes decisions using simple ideas.
This work presents a conceptual framework for causal concept-based post-hoc Explainable Artificial Intelligence (XAI), based on the requirements that explanations for non-interpretable models should be understandable as well as faithful to the model being explained. Local and global explanations are generated by calculating the probability of sufficiency of concept interventions. Example explanations are presented, generated with a proof-of-concept model made to explain classifiers trained on the CelebA dataset. Understandability is demonstrated through a clear concept-based vocabulary, subject to an implicit causal interpretation. Fidelity is addressed by highlighting important framework assumptions, stressing that the context of explanation interpretation must align with the context of explanation generation.
Similar Papers
Bridging Human Cognition and AI: A Framework for Explainable Decision-Making Systems
Statistical Finance
Makes AI easier to understand and trust.
Onto-Epistemological Analysis of AI Explanations
Artificial Intelligence
Makes AI decisions understandable and trustworthy.
From Explainable to Explanatory Artificial Intelligence: Toward a New Paradigm for Human-Centered Explanations through Generative AI
Artificial Intelligence
AI explains decisions like a helpful friend.