OMCL: Open-vocabulary Monte Carlo Localization
By: Evgenii Kruzhkov, Raphael Memmesheimer, Sven Behnke
Robust robot localization is an important prerequisite for navigation planning. If the environment map was created from different sensors, robot measurements must be robustly associated with map features. In this work, we extend Monte Carlo Localization using vision-language features. These open-vocabulary features enable to robustly compute the likelihood of visual observations, given a camera pose and a 3D map created from posed RGB-D images or aligned point clouds. The abstract vision-language features enable to associate observations and map elements from different modalities. Global localization can be initialized by natural language descriptions of the objects present in the vicinity of locations. We evaluate our approach using Matterport3D and Replica for indoor scenes and demonstrate generalization on SemanticKITTI for outdoor scenes.
Similar Papers
Robot localization in a mapped environment using Adaptive Monte Carlo algorithm
Robotics
Robot finds its way in a map.
VLG-Loc: Vision-Language Global Localization from Labeled Footprint Maps
Robotics
Helps robots find their way using landmark names.
Towards an Accurate and Effective Robot Vision (The Problem of Topological Localization for Mobile Robots)
Robotics
Helps robots know where they are using pictures.