Generative Photographic Control for Scene-Consistent Video Cinematic Editing
By: Huiqiang Sun , Liao Shen , Zhan Peng and more
Potential Business Impact:
Lets you change movie look like a pro.
Cinematic storytelling is profoundly shaped by the artful manipulation of photographic elements such as depth of field and exposure. These effects are crucial in conveying mood and creating aesthetic appeal. However, controlling these effects in generative video models remains highly challenging, as most existing methods are restricted to camera motion control. In this paper, we propose CineCtrl, the first video cinematic editing framework that provides fine control over professional camera parameters (e.g., bokeh, shutter speed). We introduce a decoupled cross-attention mechanism to disentangle camera motion from photographic inputs, allowing fine-grained, independent control without compromising scene consistency. To overcome the shortage of training data, we develop a comprehensive data generation strategy that leverages simulated photographic effects with a dedicated real-world collection pipeline, enabling the construction of a large-scale dataset for robust model training. Extensive experiments demonstrate that our model generates high-fidelity videos with precisely controlled, user-specified photographic camera effects.
Similar Papers
MotionV2V: Editing Motion in a Video
CV and Pattern Recognition
Changes how things move in videos.
BulletTime: Decoupled Control of Time and Camera Pose for Video Generation
CV and Pattern Recognition
Lets you change what happens and where the camera looks.
Light-X: Generative 4D Video Rendering with Camera and Illumination Control
CV and Pattern Recognition
Creates new videos with changing camera and light.