Multi-Object Grounding via Hierarchical Contrastive Siamese Transformers
By: Chengyi Du, Keyan Jin
Potential Business Impact:
Finds many things in 3D space from words.
Multi-object grounding in 3D scenes involves localizing multiple objects based on natural language input. While previous work has primarily focused on single-object grounding, real-world scenarios often demand the localization of several objects. To tackle this challenge, we propose Hierarchical Contrastive Siamese Transformers (H-COST), which employs a Hierarchical Processing strategy to progressively refine object localization, enhancing the understanding of complex language instructions. Additionally, we introduce a Contrastive Siamese Transformer framework, where two networks with the identical structure are used: one auxiliary network processes robust object relations from ground-truth labels to guide and enhance the second network, the reference network, which operates on segmented point-cloud data. This contrastive mechanism strengthens the model' s semantic understanding and significantly enhances its ability to process complex point-cloud data. Our approach outperforms previous state-of-the-art methods by 9.5% on challenging multi-object grounding benchmarks.
Similar Papers
Robust Visual Localization via Semantic-Guided Multi-Scale Transformer
CV and Pattern Recognition
Helps robots see where they are anywhere.
DenseGrounding: Improving Dense Language-Vision Semantics for Ego-Centric 3D Visual Grounding
CV and Pattern Recognition
Robots find objects using spoken words.
Multi-task Visual Grounding with Coarse-to-Fine Consistency Constraints
CV and Pattern Recognition
Helps computers find and outline objects in pictures.