Synchronization of Multiple Videos
By: Avihai Naaman, Ron Shapira Weber, Oren Freifeld
Potential Business Impact:
Matches up videos, even fake ones, perfectly.
Synchronizing videos captured simultaneously from multiple cameras in the same scene is often easy and typically requires only simple time shifts. However, synchronizing videos from different scenes or, more recently, generative AI videos, poses a far more complex challenge due to diverse subjects, backgrounds, and nonlinear temporal misalignment. We propose Temporal Prototype Learning (TPL), a prototype-based framework that constructs a shared, compact 1D representation from high-dimensional embeddings extracted by any of various pretrained models. TPL robustly aligns videos by learning a unified prototype sequence that anchors key action phases, thereby avoiding exhaustive pairwise matching. Our experiments show that TPL improves synchronization accuracy, efficiency, and robustness across diverse datasets, including fine-grained frame retrieval and phase classification tasks. Importantly, TPL is the first approach to mitigate synchronization issues in multiple generative AI videos depicting the same action. Our code and a new multiple video synchronization dataset are available at https://bgu-cs-vil.github.io/TPL/
Similar Papers
Visual Sync: Multi-Camera Synchronization via Cross-View Object Motion
CV and Pattern Recognition
Aligns videos from different cameras perfectly.
Beyond Audio and Pose: A General-Purpose Framework for Video Synchronization
CV and Pattern Recognition
Aligns videos from different cameras automatically.
Generative Spatiotemporal Data Augmentation
CV and Pattern Recognition
Makes computer vision work better with less data.