V-CECE: Visual Counterfactual Explanations via Conceptual Edits
By: Nikolaos Spanos , Maria Lymperaiou , Giorgos Filandrianos and more
Potential Business Impact:
Changes pictures to fool AI, showing how it thinks.
Recent black-box counterfactual generation frameworks fail to take into account the semantic content of the proposed edits, while relying heavily on training to guide the generation process. We propose a novel, plug-and-play black-box counterfactual generation framework, which suggests step-by-step edits based on theoretical guarantees of optimal edits to produce human-level counterfactual explanations with zero training. Our framework utilizes a pre-trained image editing diffusion model, and operates without access to the internals of the classifier, leading to an explainable counterfactual generation process. Throughout our experimentation, we showcase the explanatory gap between human reasoning and neural model behavior by utilizing both Convolutional Neural Network (CNN), Vision Transformer (ViT) and Large Vision Language Model (LVLM) classifiers, substantiated through a comprehensive human evaluation.
Similar Papers
DocVCE: Diffusion-based Visual Counterfactual Explanations for Document Image Classification
CV and Pattern Recognition
Shows why computers decide documents are what they are.
Towards Desiderata-Driven Design of Visual Counterfactual Explainers
Machine Learning (CS)
Shows how to change pictures to fool computers.
From Visual Explanations to Counterfactual Explanations with Latent Diffusion
CV and Pattern Recognition
Shows why computers make wrong picture guesses.