TennisTV: Do Multimodal Large Language Models Understand Tennis Rallies?
By: Zhongyuan Bao, Lejun Zhang
Potential Business Impact:
Helps computers understand fast sports like tennis.
Multimodal large language models (MLLMs) excel at general video understanding but struggle with fast, high-frequency sports like tennis, where rally clips are short yet information-dense. To systematically evaluate MLLMs in this challenging domain, we present TennisTV, the first and most comprehensive benchmark for tennis video understanding. TennisTV models each rally as a temporal-ordered sequence of consecutive stroke events, using automated pipelines for filtering and question generation. It covers 9 tasks from the stroke level to the rally level and includes 2943 human-verified questions. Evaluating 17 representative MLLMs, we provide the first systematic assessment of tennis video understanding. Results reveal substantial shortcomings and yield two key insights: (i) frame-sampling density should be tailored and balanced across tasks, and (ii) improving temporal grounding is essential for stronger reasoning.
Similar Papers
Can Large Language Models Help Multimodal Language Analysis? MMLA: A Comprehensive Benchmark
Computation and Language
Helps computers understand how people *really* talk.
Video-MMLU: A Massive Multi-Discipline Lecture Understanding Benchmark
CV and Pattern Recognition
Helps computers understand complex lectures better.
Large VLM-based Stylized Sports Captioning
CV and Pattern Recognition
Lets computers describe sports plays like a commentator.