Guidance Free Image Editing via Explicit Conditioning
By: Mehdi Noroozi , Alberto Gil Ramos , Luca Morreale and more
Potential Business Impact:
Makes AI image tools faster and better.
Current sampling mechanisms for conditional diffusion models rely mainly on Classifier Free Guidance (CFG) to generate high-quality images. However, CFG requires several denoising passes in each time step, e.g., up to three passes in image editing tasks, resulting in excessive computational costs. This paper introduces a novel conditioning technique to ease the computational burden of the well-established guidance techniques, thereby significantly improving the inference time of diffusion models. We present Explicit Conditioning (EC) of the noise distribution on the input modalities to achieve this. Intuitively, we model the noise to guide the conditional diffusion model during the diffusion process. We present evaluations on image editing tasks and demonstrate that EC outperforms CFG in generating diverse high-quality images with significantly reduced computations.
Similar Papers
CFG-EC: Error Correction Classifier-Free Guidance
Machine Learning (CS)
Makes AI pictures match your words better.
Unconditional Priors Matter! Improving Conditional Generation of Fine-Tuned Diffusion Models
CV and Pattern Recognition
Makes AI pictures and videos look better.
Learn to Guide Your Diffusion Model
Machine Learning (CS)
Makes AI pictures match words better.