Object Reconstruction under Occlusion with Generative Priors and Contact-induced Constraints
By: Minghan Zhu , Zhiyi Wang , Qihang Sun and more
Potential Business Impact:
Helps robots see and grab objects better.
Object geometry is key information for robot manipulation. Yet, object reconstruction is a challenging task because cameras only capture partial observations of objects, especially when occlusion occurs. In this paper, we leverage two extra sources of information to reduce the ambiguity of vision signals. First, generative models learn priors of the shapes of commonly seen objects, allowing us to make reasonable guesses of the unseen part of geometry. Second, contact information, which can be obtained from videos and physical interactions, provides sparse constraints on the boundary of the geometry. We combine the two sources of information through contact-guided 3D generation. The guidance formulation is inspired by drag-based editing in generative models. Experiments on synthetic and real-world data show that our approach improves the reconstruction compared to pure 3D generation and contact-based optimization.
Similar Papers
Vysics: Object Reconstruction Under Occlusion by Fusing Vision and Contact-Rich Physics
Robotics
Robot learns object shape from movement, even when hidden.
ReconViaGen: Towards Accurate Multi-view 3D Object Reconstruction via Generation
CV and Pattern Recognition
Makes 3D models from few pictures.
Open-world Hand-Object Interaction Video Generation Based on Structure and Contact-aware Representation
CV and Pattern Recognition
Makes videos of hands touching objects realistic.