GaussEdit: Adaptive 3D Scene Editing with Text and Image Prompts
By: Zhenyu Shu , Junlong Yu , Kai Chao and more
Potential Business Impact:
Changes 3D scenes with words and pictures.
This paper presents GaussEdit, a framework for adaptive 3D scene editing guided by text and image prompts. GaussEdit leverages 3D Gaussian Splatting as its backbone for scene representation, enabling convenient Region of Interest selection and efficient editing through a three-stage process. The first stage involves initializing the 3D Gaussians to ensure high-quality edits. The second stage employs an Adaptive Global-Local Optimization strategy to balance global scene coherence and detailed local edits and a category-guided regularization technique to alleviate the Janus problem. The final stage enhances the texture of the edited objects using a sophisticated image-to-image synthesis technique, ensuring that the results are visually realistic and align closely with the given prompts. Our experimental results demonstrate that GaussEdit surpasses existing methods in editing accuracy, visual fidelity, and processing speed. By successfully embedding user-specified concepts into 3D scenes, GaussEdit is a powerful tool for detailed and user-driven 3D scene editing, offering significant improvements over traditional methods.
Similar Papers
Realistic and Controllable 3D Gaussian-Guided Object Editing for Driving Video Generation
CV and Pattern Recognition
Makes self-driving cars practice tricky situations safely.
Mono4DEditor: Text-Driven 4D Scene Editing from Monocular Video via Point-Level Localization of Language-Embedded Gaussians
CV and Pattern Recognition
Changes videos based on what you type.
Enhancing non-Rigid 3D Model Deformations Using Mesh-based Gaussian Splatting
Graphics
Lets you easily bend and stretch 3D shapes.