ArtiLatent: Realistic Articulated 3D Object Generation via Structured Latents
By: Honghua Chen , Yushi Lan , Yongwei Chen and more
Potential Business Impact:
Creates realistic 3D objects that can move.
We propose ArtiLatent, a generative framework that synthesizes human-made 3D objects with fine-grained geometry, accurate articulation, and realistic appearance. Our approach jointly models part geometry and articulation dynamics by embedding sparse voxel representations and associated articulation properties, including joint type, axis, origin, range, and part category, into a unified latent space via a variational autoencoder. A latent diffusion model is then trained over this space to enable diverse yet physically plausible sampling. To reconstruct photorealistic 3D shapes, we introduce an articulation-aware Gaussian decoder that accounts for articulation-dependent visibility changes (e.g., revealing the interior of a drawer when opened). By conditioning appearance decoding on articulation state, our method assigns plausible texture features to regions that are typically occluded in static poses, significantly improving visual realism across articulation configurations. Extensive experiments on furniture-like objects from PartNet-Mobility and ACD datasets demonstrate that ArtiLatent outperforms existing approaches in geometric consistency and appearance fidelity. Our framework provides a scalable solution for articulated 3D object synthesis and manipulation.
Similar Papers
FreeArt3D: Training-Free Articulated Object Generation using 3D Diffusion
CV and Pattern Recognition
Makes robots move and bend like real things.
FreeArt3D: Training-Free Articulated Object Generation using 3D Diffusion
CV and Pattern Recognition
Makes 3D models move and look real.
UniLat3D: Geometry-Appearance Unified Latents for Single-Stage 3D Generation
CV and Pattern Recognition
Makes 3D objects from one picture fast.