Endless World: Real-Time 3D-Aware Long Video Generation
By: Ke Zhang , Yiqun Mei , Jiacong Xu and more
Potential Business Impact:
Creates endless, realistic videos that stay the same.
Producing long, coherent video sequences with stable 3D structure remains a major challenge, particularly in streaming scenarios. Motivated by this, we introduce Endless World, a real-time framework for infinite, 3D-consistent video generation.To support infinite video generation, we introduce a conditional autoregressive training strategy that aligns newly generated content with existing video frames. This design preserves long-range dependencies while remaining computationally efficient, enabling real-time inference on a single GPU without additional training overhead.Moreover, our Endless World integrates global 3D-aware attention to provide continuous geometric guidance across time. Our 3D injection mechanism enforces physical plausibility and geometric consistency throughout extended sequences, addressing key challenges in long-horizon and dynamic scene synthesis.Extensive experiments demonstrate that Endless World produces long, stable, and visually coherent videos, achieving competitive or superior performance to existing methods in both visual fidelity and spatial consistency. Our project has been available on https://bwgzk-keke.github.io/EndlessWorld/.
Similar Papers
LongVie 2: Multimodal Controllable Ultra-Long Video World Model
CV and Pattern Recognition
Makes videos that stay real and make sense.
WorldWeaver: Generating Long-Horizon Video Worlds via Rich Perception
CV and Pattern Recognition
Makes videos look real for longer without errors.
WorldGrow: Generating Infinite 3D World
CV and Pattern Recognition
Builds endless, realistic 3D worlds for games.