Score: 3

TA-Prompting: Enhancing Video Large Language Models for Dense Video Captioning via Temporal Anchors

Published: January 6, 2026 | arXiv ID: 2601.02908v1

By: Wei-Yuan Cheng , Kai-Po Chang , Chi-Pin Huang and more

BigTech Affiliations: NVIDIA

Potential Business Impact:

Helps videos tell stories with exact moments.

Business Areas:
Natural Language Processing Artificial Intelligence, Data and Analytics, Software

Dense video captioning aims to interpret and describe all temporally localized events throughout an input video. Recent state-of-the-art methods leverage large language models (LLMs) to provide detailed moment descriptions for video data. However, existing VideoLLMs remain challenging in identifying precise event boundaries in untrimmed videos, causing the generated captions to be not properly grounded. In this paper, we propose TA-Prompting, which enhances VideoLLMs via Temporal Anchors that learn to precisely localize events and prompt the VideoLLMs to perform temporal-aware video event understanding. During inference, in order to properly determine the output caption sequence from an arbitrary number of events presented within a video, we introduce an event coherent sampling strategy to select event captions with sufficient coherence across temporal events and cross-modal similarity with the given video. Through extensive experiments on benchmark datasets, we show that our TA-Prompting is favorable against state-of-the-art VideoLLMs, yielding superior performance on dense video captioning and temporal understanding tasks including moment retrieval and temporalQA.

Country of Origin
πŸ‡ΊπŸ‡Έ πŸ‡ΉπŸ‡Ό Taiwan, Province of China, United States

Page Count
17 pages

Category
Computer Science:
CV and Pattern Recognition