Holistic Explainable AI (H-XAI): Extending Transparency Beyond Developers in AI-Driven Decision Making
By: Kausik Lakkaraju, Siva Likitha Valluru, Biplav Srivastava
Potential Business Impact:
Helps everyone understand how computer decisions are made.
Current eXplainable AI (XAI) methods largely serve developers, often focusing on justifying model outputs rather than supporting diverse stakeholder needs. A recent shift toward Evaluative AI reframes explanation as a tool for hypothesis testing, but still focuses primarily on operational organizations. We introduce Holistic-XAI (H-XAI), a unified framework that integrates causal rating methods with traditional XAI methods to support explanation as an interactive, multi-method process. H-XAI allows stakeholders to ask a series of questions, test hypotheses, and compare model behavior against automatically constructed random and biased baselines. It combines instance-level and global explanations, adapting to each stakeholder's goals, whether understanding individual decisions, assessing group-level bias, or evaluating robustness under perturbations. We demonstrate the generality of our approach through two case studies spanning six scenarios: binary credit risk classification and financial time-series forecasting. H-XAI fills critical gaps left by existing XAI methods by combining causal ratings and post-hoc explanations to answer stakeholder-specific questions at both the individual decision level and the overall model level.
Similar Papers
A Comprehensive Perspective on Explainable AI across the Machine Learning Workflow
Machine Learning (CS)
Makes AI understandable and trustworthy for everyone.
From Explainable to Explanatory Artificial Intelligence: Toward a New Paradigm for Human-Centered Explanations through Generative AI
Artificial Intelligence
AI explains decisions like a helpful friend.
Unifying VXAI: A Systematic Review and Framework for the Evaluation of Explainable AI
Machine Learning (CS)
Helps AI explain its decisions clearly.