LOST-3DSG: Lightweight Open-Vocabulary 3D Scene Graphs with Semantic Tracking in Dynamic Environments
By: Sara Micol Ferraina , Michele Brienza , Francesco Argenziano and more
Potential Business Impact:
Helps robots see and follow moving things.
Tracking objects that move within dynamic environments is a core challenge in robotics. Recent research has advanced this topic significantly; however, many existing approaches remain inefficient due to their reliance on heavy foundation models. To address this limitation, we propose LOST-3DSG, a lightweight open-vocabulary 3D scene graph designed to track dynamic objects in real-world environments. Our method adopts a semantic approach to entity tracking based on word2vec and sentence embeddings, enabling an open-vocabulary representation while avoiding the necessity of storing dense CLIP visual features. As a result, LOST-3DSG achieves superior performance compared to approaches that rely on high-dimensional visual embeddings. We evaluate our method through qualitative and quantitative experiments conducted in a real 3D environment using a TIAGo robot. The results demonstrate the effectiveness and efficiency of LOST-3DSG in dynamic object tracking. Code and supplementary material are publicly available on the project website at https://lab-rococo-sapienza.github.io/lost-3dsg/.
Similar Papers
KeySG: Hierarchical Keyframe-Based 3D Scene Graphs
CV and Pattern Recognition
Helps robots understand and navigate complex places.
Towards Terrain-Aware Task-Driven 3D Scene Graph Generation in Outdoor Environments
Robotics
Helps robots understand outdoor places for better jobs.
MSGNav: Unleashing the Power of Multi-modal 3D Scene Graph for Zero-Shot Embodied Navigation
CV and Pattern Recognition
Robots learn to explore new places without practice.