Quantizing Diffusion Models from a Sampling-Aware Perspective
By: Qian Zeng , Jie Song , Yuanyu Wan and more
Potential Business Impact:
Makes AI art creation much faster and better.
Diffusion models have recently emerged as the dominant approach in visual generation tasks. However, the lengthy denoising chains and the computationally intensive noise estimation networks hinder their applicability in low-latency and resource-limited environments. Previous research has endeavored to address these limitations in a decoupled manner, utilizing either advanced samplers or efficient model quantization techniques. In this study, we uncover that quantization-induced noise disrupts directional estimation at each sampling step, further distorting the precise directional estimations of higher-order samplers when solving the sampling equations through discretized numerical methods, thereby altering the optimal sampling trajectory. To attain dual acceleration with high fidelity, we propose a sampling-aware quantization strategy, wherein a Mixed-Order Trajectory Alignment technique is devised to impose a more stringent constraint on the error bounds at each sampling step, facilitating a more linear probability flow. Extensive experiments on sparse-step fast sampling across multiple datasets demonstrate that our approach preserves the rapid convergence characteristics of high-speed samplers while maintaining superior generation quality. Code will be made publicly available soon.
Similar Papers
Diffusion Model Quantization: A Review
CV and Pattern Recognition
Makes AI art generators run on phones.
AccuQuant: Simulating Multiple Denoising Steps for Quantizing Diffusion Models
CV and Pattern Recognition
Makes AI image generators faster and smaller.
Efficient Personalization of Quantized Diffusion Model without Backpropagation
CV and Pattern Recognition
Makes AI image tools use much less computer memory.