LLM-Guided Indoor Navigation with Multimodal Map Understanding
By: Alberto Coffrini , Paolo Barsocchi , Francesco Furfari and more
Potential Business Impact:
Lets phones give directions inside buildings.
Indoor navigation presents unique challenges due to complex layouts and the unavailability of GNSS signals. Existing solutions often struggle with contextual adaptation, and typically require dedicated hardware. In this work, we explore the potential of a Large Language Model (LLM), i.e., ChatGPT, to generate natural, context-aware navigation instructions from indoor map images. We design and evaluate test cases across different real-world environments, analyzing the effectiveness of LLMs in interpreting spatial layouts, handling user constraints, and planning efficient routes. Our findings demonstrate the potential of LLMs for supporting personalized indoor navigation, with an average of 86.59% correct indications and a maximum of 97.14%. The proposed system achieves high accuracy and reasoning performance. These results have key implications for AI-driven navigation and assistive technologies.
Similar Papers
Research on Navigation Methods Based on LLMs
Robotics
Lets robots find their way around buildings.
Vision-Based Localization and LLM-based Navigation for Indoor Environments
Machine Learning (CS)
Guides you indoors using phone camera and AI.
GeoNav: Empowering MLLMs with Explicit Geospatial Reasoning Abilities for Language-Goal Aerial Navigation
Robotics
Drones find places using words and maps.