Noise-Level Diffusion Guidance: Well Begun is Half Done
By: Harvey Mannering, Zhiwu Huang, Adam Prugel-Bennett
Potential Business Impact:
Makes AI pictures better and more like you ask.
Diffusion models have achieved state-of-the-art image generation. However, the random Gaussian noise used to start the diffusion process influences the final output, causing variations in image quality and prompt adherence. Existing noise-level optimization approaches generally rely on extra dataset construction, additional networks, or backpropagation-based optimization, limiting their practicality. In this paper, we propose Noise Level Guidance (NLG), a simple, efficient, and general noise-level optimization approach that refines initial noise by increasing the likelihood of its alignment with general guidance - requiring no additional training data, auxiliary networks, or backpropagation. The proposed NLG approach provides a unified framework generalizable to both conditional and unconditional diffusion models, accommodating various forms of diffusion-level guidance. Extensive experiments on five standard benchmarks demonstrate that our approach enhances output generation quality and input condition adherence. By seamlessly integrating with existing guidance methods while maintaining computational efficiency, our method establishes NLG as a practical and scalable enhancement to diffusion models. Code can be found at https://github.com/harveymannering/NoiseLevelGuidance.
Similar Papers
GuidNoise: Single-Pair Guided Diffusion for Generalized Noise Synthesis
CV and Pattern Recognition
Makes blurry photos clear with less work.
Mitigating the Noise Shift for Denoising Generative Models via Noise Awareness Guidance
Machine Learning (CS)
Fixes AI image makers to create better pictures.
Noise Projection: Closing the Prompt-Agnostic Gap Behind Text-to-Image Misalignment in Diffusion Models
CV and Pattern Recognition
Makes AI pictures match words better.