Equilibrium Matching: Generative Modeling with Implicit Energy-Based Models
By: Runqian Wang, Yilun Du
Potential Business Impact:
Makes computers create realistic pictures faster.
We introduce Equilibrium Matching (EqM), a generative modeling framework built from an equilibrium dynamics perspective. EqM discards the non-equilibrium, time-conditional dynamics in traditional diffusion and flow-based generative models and instead learns the equilibrium gradient of an implicit energy landscape. Through this approach, we can adopt an optimization-based sampling process at inference time, where samples are obtained by gradient descent on the learned landscape with adjustable step sizes, adaptive optimizers, and adaptive compute. EqM surpasses the generation performance of diffusion/flow models empirically, achieving an FID of 1.90 on ImageNet 256$\times$256. EqM is also theoretically justified to learn and sample from the data manifold. Beyond generation, EqM is a flexible framework that naturally handles tasks including partially noised image denoising, OOD detection, and image composition. By replacing time-conditional velocities with a unified equilibrium landscape, EqM offers a tighter bridge between flow and energy-based models and a simple route to optimization-driven inference.
Similar Papers
Energy Matching: Unifying Flow Matching and Energy-Based Models for Generative Modeling
Machine Learning (CS)
Makes AI create better pictures with more control.
Energy-Weighted Flow Matching for Offline Reinforcement Learning
Machine Learning (CS)
Makes AI learn better from past information.
On the flow matching interpretability
Machine Learning (CS)
Makes AI understand how things change physically.