Training Feature Attribution for Vision Models
By: Aziz Bacha, Thomas George
Potential Business Impact:
Shows how bad training pictures trick computers.
Deep neural networks are often considered opaque systems, prompting the need for explainability methods to improve trust and accountability. Existing approaches typically attribute test-time predictions either to input features (e.g., pixels in an image) or to influential training examples. We argue that both perspectives should be studied jointly. This work explores *training feature attribution*, which links test predictions to specific regions of specific training images and thereby provides new insights into the inner workings of deep models. Our experiments on vision datasets show that training feature attribution yields fine-grained, test-specific explanations: it identifies harmful examples that drive misclassifications and reveals spurious correlations, such as patch-based shortcuts, that conventional attribution methods fail to expose.
Similar Papers
Attribution Explanations for Deep Neural Networks: A Theoretical Perspective
Machine Learning (CS)
Makes AI decisions easier to understand.
Distribution-Based Feature Attribution for Explaining the Predictions of Any Classifier
Machine Learning (CS)
Explains AI decisions using data patterns.
Higher-Order Feature Attribution: Bridging Statistics, Explainable AI, and Topological Signal Processing
Machine Learning (CS)
Explains how computer decisions are made.