CounterVQA: Evaluating and Improving Counterfactual Reasoning in Vision-Language Models for Video Understanding
By: Yuefei Chen , Jiang Liu , Xiaodong Lin and more
Potential Business Impact:
Helps computers imagine "what if" in videos.
Vision Language Models (VLMs) have recently shown significant advancements in video understanding, especially in feature alignment, event reasoning, and instruction-following tasks. However, their capability for counterfactual reasoning, inferring alternative outcomes under hypothetical conditions, remains underexplored. This capability is essential for robust video understanding, as it requires identifying underlying causal structures and reasoning about unobserved possibilities, rather than merely recognizing observed patterns. To systematically evaluate this capability, we introduce CounterVQA, a video-based benchmark featuring three progressive difficulty levels that assess different aspects of counterfactual reasoning. Through comprehensive evaluation of both state-of-the-art open-source and closed-source models, we uncover a substantial performance gap: while these models achieve reasonable accuracy on simple counterfactual questions, performance degrades significantly on complex multi-hop causal chains. To address these limitations, we develop a post-training method, CFGPT, that enhances a model's visual counterfactual reasoning ability by distilling its counterfactual reasoning capability from the language modality, yielding consistent improvements across all CounterVQA difficulty levels. Dataset and code will be further released.
Similar Papers
What's Missing in Vision-Language Models? Probing Their Struggles with Causal Order Reasoning
Computation and Language
Helps computers understand why things happen.
CausalVLBench: Benchmarking Visual Causal Reasoning in Large Vision-Language Models
Machine Learning (CS)
Helps computers understand cause and effect in pictures.
Beyond Generation: Multi-Hop Reasoning for Factual Accuracy in Vision-Language Models
Artificial Intelligence
Makes AI understand pictures and facts better.