Learning to Generate 4D LiDAR Sequences
By: Ao Liang , Youquan Liu , Yu Yang and more
Potential Business Impact:
Creates 3D car sensor data from words.
While generative world models have advanced video and occupancy-based data synthesis, LiDAR generation remains underexplored despite its importance for accurate 3D perception. Extending generation to 4D LiDAR data introduces challenges in controllability, temporal stability, and evaluation. We present LiDARCrafter, a unified framework that converts free-form language into editable LiDAR sequences. Instructions are parsed into ego-centric scene graphs, which a tri-branch diffusion model transforms into object layouts, trajectories, and shapes. A range-image diffusion model generates the initial scan, and an autoregressive module extends it into a temporally coherent sequence. The explicit layout design further supports object-level editing, such as insertion or relocation. To enable fair assessment, we provide EvalSuite, a benchmark spanning scene-, object-, and sequence-level metrics. On nuScenes, LiDARCrafter achieves state-of-the-art fidelity, controllability, and temporal consistency, offering a foundation for LiDAR-based simulation and data augmentation.
Similar Papers
LiDARCrafter: Dynamic 4D World Modeling from LiDAR Sequences
CV and Pattern Recognition
Makes self-driving cars "see" and move better.
LiDARCrafter: Dynamic 4D World Modeling from LiDAR Sequences
CV and Pattern Recognition
Makes self-driving cars "see" and change scenes.
DriveLiDAR4D: Sequential and Controllable LiDAR Scene Generation for Autonomous Driving
CV and Pattern Recognition
Creates realistic driving scenes for self-driving cars.