Dynamic Camera Poses and Where to Find Them
By: Chris Rockwell , Joseph Tung , Tsung-Yi Lin and more
Potential Business Impact:
Creates realistic videos by tracking camera movement.
Annotating camera poses on dynamic Internet videos at scale is critical for advancing fields like realistic video generation and simulation. However, collecting such a dataset is difficult, as most Internet videos are unsuitable for pose estimation. Furthermore, annotating dynamic Internet videos present significant challenges even for state-of-theart methods. In this paper, we introduce DynPose-100K, a large-scale dataset of dynamic Internet videos annotated with camera poses. Our collection pipeline addresses filtering using a carefully combined set of task-specific and generalist models. For pose estimation, we combine the latest techniques of point tracking, dynamic masking, and structure-from-motion to achieve improvements over the state-of-the-art approaches. Our analysis and experiments demonstrate that DynPose-100K is both large-scale and diverse across several key attributes, opening up avenues for advancements in various downstream applications.
Similar Papers
DynOPETs: A Versatile Benchmark for Dynamic Object Pose Estimation and Tracking in Moving Camera Scenarios
CV and Pattern Recognition
Helps robots see moving things from moving cameras.
InterPose: Learning to Generate Human-Object Interactions from Large-Scale Web Videos
CV and Pattern Recognition
Makes computer characters interact with objects realistically.
6D Object Pose Tracking in Internet Videos for Robotic Manipulation
CV and Pattern Recognition
Robots copy object moves from online videos.