A.I.R.: Enabling Adaptive, Iterative, and Reasoning-based Frame Selection For Video Question Answering
By: Yuanhao Zou , Shengji Jin , Andong Deng and more
Potential Business Impact:
Helps AI understand videos by picking key moments.
Effectively applying Vision-Language Models (VLMs) to Video Question Answering (VideoQA) hinges on selecting a concise yet comprehensive set of frames, as processing entire videos is computationally infeasible. However, current frame selection methods face a critical trade-off: approaches relying on lightweight similarity models, such as CLIP, often fail to capture the nuances of complex queries, resulting in inaccurate similarity scores that cannot reflect the authentic query-frame relevance, which further undermines frame selection. Meanwhile, methods that leverage a VLM for deeper analysis achieve higher accuracy but incur prohibitive computational costs. To address these limitations, we propose A.I.R., a training-free approach for Adaptive, Iterative, and Reasoning-based frame selection. We leverage a powerful VLM to perform deep, semantic analysis on complex queries, and this analysis is deployed within a cost-effective iterative loop that processes only a small batch of the most high-potential frames at a time. Extensive experiments on various VideoQA benchmarks demonstrate that our approach outperforms existing frame selection methods, significantly boosts the performance of the foundation VLM, and achieves substantial gains in computational efficiency over other VLM-based techniques.
Similar Papers
Improving Video Question Answering through query-based frame selection
CV and Pattern Recognition
Helps computers understand videos better by picking important scenes.
POVQA: Preference-Optimized Video Question Answering with Rationales for Data Efficiency
CV and Pattern Recognition
Lets computers understand long videos by summarizing them.
ReasVQA: Advancing VideoQA with Imperfect Reasoning Process
CV and Pattern Recognition
Helps computers understand videos to answer questions.