Score: 1

World Models for Autonomous Navigation of Terrestrial Robots from LIDAR Observations

Published: December 3, 2025 | arXiv ID: 2512.03429v1

By: Raul Steinmetz , Fabio Demo Rosa , Victor Augusto Kich and more

Potential Business Impact:

Helps robots learn to drive better using less data.

Business Areas:
Autonomous Vehicles Transportation

Autonomous navigation of terrestrial robots using Reinforcement Learning (RL) from LIDAR observations remains challenging due to the high dimensionality of sensor data and the sample inefficiency of model-free approaches. Conventional policy networks struggle to process full-resolution LIDAR inputs, forcing prior works to rely on simplified observations that reduce spatial awareness and navigation robustness. This paper presents a novel model-based RL framework built on top of the DreamerV3 algorithm, integrating a Multi-Layer Perceptron Variational Autoencoder (MLP-VAE) within a world model to encode high-dimensional LIDAR readings into compact latent representations. These latent features, combined with a learned dynamics predictor, enable efficient imagination-based policy optimization. Experiments on simulated TurtleBot3 navigation tasks demonstrate that the proposed architecture achieves faster convergence and higher success rate compared to model-free baselines such as SAC, DDPG, and TD3. It is worth emphasizing that the DreamerV3-based agent attains a 100% success rate across all evaluated environments when using the full dataset of the Turtlebot3 LIDAR (360 readings), while model-free methods plateaued below 85%. These findings demonstrate that integrating predictive world models with learned latent representations enables more efficient and robust navigation from high-dimensional sensory data.

Country of Origin
🇧🇷 Brazil

Repos / Data Links

Page Count
10 pages

Category
Computer Science:
Robotics