MOSU: Autonomous Long-range Robot Navigation with Multi-modal Scene Understanding
By: Jing Liang , Kasun Weerakoon , Daeun Song and more
Potential Business Impact:
Helps robots drive safely on roads.
We present MOSU, a novel autonomous long-range navigation system that enhances global navigation for mobile robots through multimodal perception and on-road scene understanding. MOSU addresses the outdoor robot navigation challenge by integrating geometric, semantic, and contextual information to ensure comprehensive scene understanding. The system combines GPS and QGIS map-based routing for high-level global path planning and multi-modal trajectory generation for local navigation refinement. For trajectory generation, MOSU leverages multi-modalities: LiDAR-based geometric data for precise obstacle avoidance, image-based semantic segmentation for traversability assessment, and Vision-Language Models (VLMs) to capture social context and enable the robot to adhere to social norms in complex environments. This multi-modal integration improves scene understanding and enhances traversability, allowing the robot to adapt to diverse outdoor conditions. We evaluate our system in real-world on-road environments and benchmark it on the GND dataset, achieving a 10% improvement in traversability on navigable terrains while maintaining a comparable navigation distance to existing global navigation methods.
Similar Papers
MUSON: A Reasoning-oriented Multimodal Dataset for Socially Compliant Navigation in Urban Environments
CV and Pattern Recognition
Helps robots safely walk through crowds.
Enhancing Situational Awareness in Underwater Robotics with Multi-modal Spatial Perception
Robotics
Helps robots see and map underwater clearly.
MSGNav: Unleashing the Power of Multi-modal 3D Scene Graph for Zero-Shot Embodied Navigation
CV and Pattern Recognition
Robots learn to explore new places without practice.