Can Image-To-Video Models Simulate Pedestrian Dynamics?
By: Aaron Appelle, Jerome P. Lynch
Potential Business Impact:
Makes videos of people walking realistically.
Recent high-performing image-to-video (I2V) models based on variants of the diffusion transformer (DiT) have displayed remarkable inherent world-modeling capabilities by virtue of training on large scale video datasets. We investigate whether these models can generate realistic pedestrian movement patterns in crowded public scenes. Our framework conditions I2V models on keyframes extracted from pedestrian trajectory benchmarks, then evaluates their trajectory prediction performance using quantitative measures of pedestrian dynamics.
Similar Papers
Evaluating Video Models as Simulators of Multi-Person Pedestrian Trajectories
CV and Pattern Recognition
Makes computer videos show people walking realistically.
MobileI2V: Fast and High-Resolution Image-to-Video on Mobile Devices
CV and Pattern Recognition
Makes videos from pictures on phones fast.
Extrapolating and Decoupling Image-to-Video Generation Models: Motion Modeling is Easier Than You Think
CV and Pattern Recognition
Makes still pictures move with text instructions.