Causality Model for Semantic Understanding on Videos
By: Li Yicong
Potential Business Impact:
Teaches computers to truly understand videos.
After a decade of prosperity, the development of video understanding has reached a critical juncture, where the sole reliance on massive data and complex architectures is no longer a one-size-fits-all solution to all situations. The presence of ubiquitous data imbalance hampers DNNs from effectively learning the underlying causal mechanisms, leading to significant performance drops when encountering distribution shifts, such as long-tail imbalances and perturbed imbalances. This realization has prompted researchers to seek alternative methodologies to capture causal patterns in video data. To tackle these challenges and increase the robustness of DNNs, causal modeling emerged as a principle to discover the true causal patterns behind the observed correlations. This thesis focuses on the domain of semantic video understanding and explores the potential of causal modeling to advance two fundamental tasks: Video Relation Detection (VidVRD) and Video Question Answering (VideoQA).
Similar Papers
CausalVQA: A Physically Grounded Causal Reasoning Benchmark for Video Models
CV and Pattern Recognition
Teaches computers to understand cause and effect.
Admitting Ignorance Helps the Video Question Answering Models to Answer
CV and Pattern Recognition
Teaches computers to admit when they don't know answers.
VersaVid-R1: A Versatile Video Understanding and Reasoning Model from Question Answering to Captioning Tasks
CV and Pattern Recognition
Helps computers understand and answer questions about videos.