IntrinsicEdit: Precise generative image manipulation in intrinsic space
By: Linjie Lyu , Valentin Deschaintre , Yannick Hold-Geoffroy and more
Potential Business Impact:
Edits pictures exactly how you want.
Generative diffusion models have advanced image editing with high-quality results and intuitive interfaces such as prompts and semantic drawing. However, these interfaces lack precise control, and the associated methods typically specialize on a single editing task. We introduce a versatile, generative workflow that operates in an intrinsic-image latent space, enabling semantic, local manipulation with pixel precision for a range of editing operations. Building atop the RGB-X diffusion framework, we address key challenges of identity preservation and intrinsic-channel entanglement. By incorporating exact diffusion inversion and disentangled channel manipulation, we enable precise, efficient editing with automatic resolution of global illumination effects -- all without additional data collection or model fine-tuning. We demonstrate state-of-the-art performance across a variety of tasks on complex images, including color and texture adjustments, object insertion and removal, global relighting, and their combinations.
Similar Papers
Edit2Perceive: Image Editing Diffusion Models Are Strong Dense Perceivers
CV and Pattern Recognition
Makes computers understand pictures better for tasks.
Towards Spatially Consistent Image Generation: On Incorporating Intrinsic Scene Properties into Diffusion Models
CV and Pattern Recognition
Makes AI pictures look more real and organized.
Energy-Guided Optimization for Personalized Image Editing with Pretrained Text-to-Image Diffusion Models
CV and Pattern Recognition
Changes pictures to match your exact ideas.