Toward Faithfulness-guided Ensemble Interpretation of Neural Network
By: Siyu Zhang, Kenneth Mcmillan
Potential Business Impact:
Shows how computer brains make decisions clearly.
Interpretable and faithful explanations for specific neural inferences are crucial for understanding and evaluating model behavior. Our work introduces \textbf{F}aithfulness-guided \textbf{E}nsemble \textbf{I}nterpretation (\textbf{FEI}), an innovative framework that enhances the breadth and effectiveness of faithfulness, advancing interpretability by providing superior visualization. Through an analysis of existing evaluation benchmarks, \textbf{FEI} employs a smooth approximation to elevate quantitative faithfulness scores. Diverse variations of \textbf{FEI} target enhanced faithfulness in hidden layer encodings, expanding interpretability. Additionally, we propose a novel qualitative metric that assesses hidden layer faithfulness. In extensive experiments, \textbf{FEI} surpasses existing methods, demonstrating substantial advances in qualitative visualization and quantitative faithfulness scores. Our research establishes a comprehensive framework for elevating faithfulness in neural network explanations, emphasizing both breadth and precision
Similar Papers
FaCT: Faithful Concept Traces for Explaining Neural Network Decisions
Machine Learning (CS)
Explains how computer "brains" understand pictures.
DeepFaith: A Domain-Free and Model-Agnostic Unified Framework for Highly Faithful Explanations
Machine Learning (CS)
Makes AI explain its choices clearly.
Interpretive Efficiency: Information-Geometric Foundations of Data Usefulness
Machine Learning (CS)
Measures how well AI understands what it sees.