Enhancing Interpretability for Vision Models via Shapley Value Optimization
By: Kanglong Fan, Yunqiao Yang, Chen Ma
Potential Business Impact:
Explains how computers make choices, clearly.
Deep neural networks have demonstrated remarkable performance across various domains, yet their decision-making processes remain opaque. Although many explanation methods are dedicated to bringing the obscurity of DNNs to light, they exhibit significant limitations: post-hoc explanation methods often struggle to faithfully reflect model behaviors, while self-explaining neural networks sacrifice performance and compatibility due to their specialized architectural designs. To address these challenges, we propose a novel self-explaining framework that integrates Shapley value estimation as an auxiliary task during training, which achieves two key advancements: 1) a fair allocation of the model prediction scores to image patches, ensuring explanations inherently align with the model's decision logic, and 2) enhanced interpretability with minor structural modifications, preserving model performance and compatibility. Extensive experiments on multiple benchmarks demonstrate that our method achieves state-of-the-art interpretability.
Similar Papers
Clinical Interpretability of Deep Learning Segmentation Through Shapley-Derived Agreement and Uncertainty Metrics
Image and Video Processing
Helps doctors trust AI that finds sickness in scans.
Fast-DataShapley: Neural Modeling for Training Data Valuation
Machine Learning (CS)
Rewards data creators fairly and fast for AI.
Here Comes the Explanation: A Shapley Perspective on Multi-contrast Medical Image Segmentation
Image and Video Processing
Helps doctors understand how AI finds brain tumors.