Follow the Signs: Using Textual Cues and LLMs to Guide Efficient Robot Navigation
By: Jing Cao, Nishanth Kumar, Aidan Curtis
Potential Business Impact:
Helps robots find rooms using text clues.
Autonomous navigation in unfamiliar environments often relies on geometric mapping and planning strategies that overlook rich semantic cues such as signs, room numbers, and textual labels. We propose a novel semantic navigation framework that leverages large language models (LLMs) to infer patterns from partial observations and predict regions where the goal is most likely located. Our method combines local perceptual inputs with frontier-based exploration and periodic LLM queries, which extract symbolic patterns (e.g., room numbering schemes and building layout structures) and update a confidence grid used to guide exploration. This enables robots to move efficiently toward goal locations labeled with textual identifiers (e.g., "room 8") even before direct observation. We demonstrate that this approach enables more efficient navigation in sparse, partially observable grid environments by exploiting symbolic patterns. Experiments across environments modeled after real floor plans show that our approach consistently achieves near-optimal paths and outperforms baselines by over 25% in Success weighted by Path Length.
Similar Papers
Bootstrapped LLM Semantics for Context-Aware Path Planning
Robotics
Robots learn to move safely using language.
Research on Navigation Methods Based on LLMs
Robotics
Lets robots find their way around buildings.
LLM-Guided Indoor Navigation with Multimodal Map Understanding
Artificial Intelligence
Lets phones give directions inside buildings.