Over++: Generative Video Compositing for Layer Interaction Effects
By: Luchao Qi , Jiaye Wu , Jun Myeong Choi and more
In professional video compositing workflows, artists must manually create environmental interactions-such as shadows, reflections, dust, and splashes-between foreground subjects and background layers. Existing video generative models struggle to preserve the input video while adding such effects, and current video inpainting methods either require costly per-frame masks or yield implausible results. We introduce augmented compositing, a new task that synthesizes realistic, semi-transparent environmental effects conditioned on text prompts and input video layers, while preserving the original scene. To address this task, we present Over++, a video effect generation framework that makes no assumptions about camera pose, scene stationarity, or depth supervision. We construct a paired effect dataset tailored for this task and introduce an unpaired augmentation strategy that preserves text-driven editability. Our method also supports optional mask control and keyframe guidance without requiring dense annotations. Despite training on limited data, Over++ produces diverse and realistic environmental effects and outperforms existing baselines in both effect generation and scene preservation.
Similar Papers
GenCompositor: Generative Video Compositing with Diffusion Transformer
CV and Pattern Recognition
Lets you easily add video clips into movies.
Omni-Effects: Unified and Spatially-Controllable Visual Effects Generation
CV and Pattern Recognition
Combines many movie effects in one place.
Generative Photographic Control for Scene-Consistent Video Cinematic Editing
CV and Pattern Recognition
Lets you change movie look like a pro.