Floorplan2Guide: LLM-Guided Floorplan Parsing for BLV Indoor Navigation
By: Aydin Ayanzadeh, Tim Oates
Indoor navigation remains a critical challenge for people with visual impairments. The current solutions mainly rely on infrastructure-based systems, which limit their ability to navigate safely in dynamic environments. We propose a novel navigation approach that utilizes a foundation model to transform floor plans into navigable knowledge graphs and generate human-readable navigation instructions. Floorplan2Guide integrates a large language model (LLM) to extract spatial information from architectural layouts, reducing the manual preprocessing required by earlier floorplan parsing methods. Experimental results indicate that few-shot learning improves navigation accuracy in comparison to zero-shot learning on simulated and real-world evaluations. Claude 3.7 Sonnet achieves the highest accuracy among the evaluated models, with 92.31%, 76.92%, and 61.54% on the short, medium, and long routes, respectively, under 5-shot prompting of the MP-1 floor plan. The success rate of graph-based spatial structure is 15.4% higher than that of direct visual reasoning among all models, which confirms that graphical representation and in-context learning enhance navigation performance and make our solution more precise for indoor navigation of Blind and Low Vision (BLV) users.
Similar Papers
LLM-Guided Indoor Navigation with Multimodal Map Understanding
Artificial Intelligence
Lets phones give directions inside buildings.
Vision-Based Localization and LLM-based Navigation for Indoor Environments
Machine Learning (CS)
Guides you indoors using phone camera and AI.
Floor Plan-Guided Visual Navigation Incorporating Depth and Directional Cues
Robotics
Helps robots find their way using pictures and maps.