Embodied Spatial Intelligence: from Implicit Scene Modeling to Spatial Reasoning
By: Jiading Fang
Potential Business Impact:
Robots understand and follow spoken directions.
This thesis introduces "Embodied Spatial Intelligence" to address the challenge of creating robots that can perceive and act in the real world based on natural language instructions. To bridge the gap between Large Language Models (LLMs) and physical embodiment, we present contributions on two fronts: scene representation and spatial reasoning. For perception, we develop robust, scalable, and accurate scene representations using implicit neural models, with contributions in self-supervised camera calibration, high-fidelity depth field generation, and large-scale reconstruction. For spatial reasoning, we enhance the spatial capabilities of LLMs by introducing a novel navigation benchmark, a method for grounding language in 3D, and a state-feedback mechanism to improve long-horizon decision-making. This work lays a foundation for robots that can robustly perceive their surroundings and intelligently act upon complex, language-based commands.
Similar Papers
A Survey of Large Language Model-Powered Spatial Intelligence Across Scales: Advances in Embodied Agents, Smart Cities, and Earth Science
Artificial Intelligence
Helps computers understand and use space better.
From reactive to cognitive: brain-inspired spatial intelligence for embodied agents
Artificial Intelligence
Helps robots learn and remember places like humans.
Embodied-R: Collaborative Framework for Activating Embodied Spatial Reasoning in Foundation Models via Reinforcement Learning
Artificial Intelligence
Helps computers understand space by watching videos.