Cost-Sensitive Learning for Long-Tailed Temporal Action Segmentation
By: Zhanzhong Pang , Fadime Sener , Shrinivas Ramasubramanian and more
Potential Business Impact:
Helps videos understand actions, even rare ones.
Temporal action segmentation in untrimmed procedural videos aims to densely label frames into action classes. These videos inherently exhibit long-tailed distributions, where actions vary widely in frequency and duration. In temporal action segmentation approaches, we identified a bi-level learning bias. This bias encompasses (1) a class-level bias, stemming from class imbalance favoring head classes, and (2) a transition-level bias arising from variations in transitions, prioritizing commonly observed transitions. As a remedy, we introduce a constrained optimization problem to alleviate both biases. We define learning states for action classes and their associated transitions and integrate them into the optimization process. We propose a novel cost-sensitive loss function formulated as a weighted cross-entropy loss, with weights adaptively adjusted based on the learning state of actions and their transitions. Experiments on three challenging temporal segmentation benchmarks and various frameworks demonstrate the effectiveness of our approach, resulting in significant improvements in both per-class frame-wise and segment-wise performance.
Similar Papers
Pose-Aware Weakly-Supervised Action Segmentation
CV and Pattern Recognition
Teaches computers to understand actions in videos.
Towards Precise Action Spotting: Addressing Temporal Misalignment in Labels with Dynamic Label Assignment
CV and Pattern Recognition
Fixes computer action spotting when labels are messy.
Towards Generalizing Temporal Action Segmentation to Unseen Views
CV and Pattern Recognition
Helps videos understand actions from new angles.