Can LLMs Learn to Map the World from Local Descriptions?
By: Sirui Xia , Aili Chen , Xintao Wang and more
Potential Business Impact:
Helps computers understand maps and find directions.
Recent advances in Large Language Models (LLMs) have demonstrated strong capabilities in tasks such as code and mathematics. However, their potential to internalize structured spatial knowledge remains underexplored. This study investigates whether LLMs, grounded in locally relative human observations, can construct coherent global spatial cognition by integrating fragmented relational descriptions. We focus on two core aspects of spatial cognition: spatial perception, where models infer consistent global layouts from local positional relationships, and spatial navigation, where models learn road connectivity from trajectory data and plan optimal paths between unconnected locations. Experiments conducted in a simulated urban environment demonstrate that LLMs not only generalize to unseen spatial relationships between points of interest (POIs) but also exhibit latent representations aligned with real-world spatial distributions. Furthermore, LLMs can learn road connectivity from trajectory descriptions, enabling accurate path planning and dynamic spatial awareness during navigation.
Similar Papers
Can Large Language Models Integrate Spatial Data? Empirical Insights into Reasoning Strengths and Computational Weaknesses
Artificial Intelligence
Helps computers combine messy map data better.
Linear Spatial World Models Emerge in Large Language Models
Artificial Intelligence
Computers learn how objects are arranged in space.
From Text to Space: Mapping Abstract Spatial Models in LLMs during a Grid-World Navigation Task
Artificial Intelligence
Helps AI understand maps for better navigation.