Decoupled MeanFlow: Turning Flow Models into Flow Maps for Accelerated Sampling
By: Kyungmin Lee, Sihyun Yu, Jinwoo Shin
Potential Business Impact:
Makes AI create pictures much faster.
Denoising generative models, such as diffusion and flow-based models, produce high-quality samples but require many denoising steps due to discretization error. Flow maps, which estimate the average velocity between timesteps, mitigate this error and enable faster sampling. However, their training typically demands architectural changes that limit compatibility with pretrained flow models. We introduce Decoupled MeanFlow, a simple decoding strategy that converts flow models into flow map models without architectural modifications. Our method conditions the final blocks of diffusion transformers on the subsequent timestep, allowing pretrained flow models to be directly repurposed as flow maps. Combined with enhanced training techniques, this design enables high-quality generation in as few as 1 to 4 steps. Notably, we find that training flow models and subsequently converting them is more efficient and effective than training flow maps from scratch. On ImageNet 256x256 and 512x512, our models attain 1-step FID of 2.16 and 2.12, respectively, surpassing prior art by a large margin. Furthermore, we achieve FID of 1.51 and 1.68 when increasing the steps to 4, which nearly matches the performance of flow models while delivering over 100x faster inference.
Similar Papers
From Diffusion to One-Step Generation: A Comparative Study of Flow-Based Models with Application to Image Inpainting
CV and Pattern Recognition
Makes pictures from scratch in one step.
Improved Mean Flows: On the Challenges of Fastforward Generative Models
CV and Pattern Recognition
Makes AI create pictures faster and better.
Flow Straighter and Faster: Efficient One-Step Generative Modeling via MeanFlow on Rectified Trajectories
CV and Pattern Recognition
Makes AI create images much faster and better.