CoFE: A Framework Generating Counterfactual ECG for Explainable Cardiac AI-Diagnostics
By: Jong-Hwan Jang, Junho Song, Yong-Yeon Jo
Potential Business Impact:
Shows how heart signals help doctors decide.
Recognizing the need for explainable AI (XAI) approaches to enable the successful integration of AI-based ECG prediction models (AI-ECG) into clinical practice, we introduce a framework generating \textbf{Co}unter\textbf{F}actual \textbf{E}CGs (i,e., named CoFE) to illustrate how specific features, such as amplitudes and intervals, influence the model's predictive decisions. To demonstrate the applicability of the CoFE, we present two case studies: atrial fibrillation classification and potassium level regression models. The CoFE reveals feature changes in ECG signals that align with the established clinical knowledge. By clarifying both \textbf{where valid features appear} in the ECG and \textbf{how they influence the model's predictions}, we anticipate that our framework will enhance the interpretability of AI-ECG models and support more effective clinical decision-making. Our demonstration video is available at: https://www.youtube.com/watch?v=YoW0bNBPglQ.
Similar Papers
From Prototypes to Sparse ECG Explanations: SHAP-Driven Counterfactuals for Multivariate Time-Series Multi-class Classification
Machine Learning (CS)
Explains heart monitor results by showing what to change.
ACE: Adapting sampling for Counterfactual Explanations
Machine Learning (CS)
Finds smallest changes to fix computer guesses.
Explainable AI (XAI) for Arrhythmia detection from electrocardiograms
Machine Learning (CS)
Shows doctors why heart rhythm is wrong.