NavAI: A Generalizable LLM Framework for Navigation Tasks in Virtual Reality Environments
By: Xue Qin, Matthew DiGiovanni
Potential Business Impact:
Helps VR characters find their way around.
Navigation is one of the fundamental tasks for automated exploration in Virtual Reality (VR). Existing technologies primarily focus on path optimization in 360-degree image datasets and 3D simulators, which cannot be directly applied to immersive VR environments. To address this gap, we present NavAI, a generalizable large language model (LLM)-based navigation framework that supports both basic actions and complex goal-directed tasks across diverse VR applications. We evaluate NavAI in three distinct VR environments through goal-oriented and exploratory tasks. Results show that it achieves high accuracy, with an 89% success rate in goal-oriented tasks. Our analysis also highlights current limitations of relying entirely on LLMs, particularly in scenarios that require dynamic goal assessment. Finally, we discuss the limitations observed during the experiments and offer insights for future research directions.
Similar Papers
IndustryNav: Exploring Spatial Reasoning of Embodied Agents in Dynamic Industrial Navigation
Robotics
Helps robots safely navigate busy factories.
VL-Nav: Real-time Vision-Language Navigation with Spatial Reasoning
Robotics
Robot finds things using words and seeing.
ImagineNav++: Prompting Vision-Language Models as Embodied Navigator through Scene Imagination
Robotics
Robots learn to explore homes by imagining where to go.