Efficient Zero-Shot Inpainting with Decoupled Diffusion Guidance
By: Badr Moufad , Navid Bagheri Shouraki , Alain Oliviero Durmus and more
Diffusion models have emerged as powerful priors for image editing tasks such as inpainting and local modification, where the objective is to generate realistic content that remains consistent with observed regions. In particular, zero-shot approaches that leverage a pretrained diffusion model, without any retraining, have been shown to achieve highly effective reconstructions. However, state-of-the-art zero-shot methods typically rely on a sequence of surrogate likelihood functions, whose scores are used as proxies for the ideal score. This procedure however requires vector-Jacobian products through the denoiser at every reverse step, introducing significant memory and runtime overhead. To address this issue, we propose a new likelihood surrogate that yields simple and efficient to sample Gaussian posterior transitions, sidestepping the backpropagation through the denoiser network. Our extensive experiments show that our method achieves strong observation consistency compared with fine-tuned baselines and produces coherent, high-quality reconstructions, all while significantly reducing inference cost. Code is available at https://github.com/YazidJanati/ding.
Similar Papers
Zero-Shot Solving of Imaging Inverse Problems via Noise-Refined Likelihood Guided Diffusion Models
CV and Pattern Recognition
Fixes blurry pictures without retraining.
Controllable Localized Face Anonymization Via Diffusion Inpainting
CV and Pattern Recognition
Hides faces in pictures while keeping them useful.
Zero-Shot Video Deraining with Video Diffusion Models
CV and Pattern Recognition
Clears rain from videos without needing special training.