Incorporating Inductive Biases to Energy-based Generative Models
By: Yukun Li, Li-Ping Liu
Potential Business Impact:
Helps computers learn and create better by understanding data.
With the advent of score-matching techniques for model training and Langevin dynamics for sample generation, energy-based models (EBMs) have gained renewed interest as generative models. Recent EBMs usually use neural networks to define their energy functions. In this work, we introduce a novel hybrid approach that combines an EBM with an exponential family model to incorporate inductive bias into data modeling. Specifically, we augment the energy term with a parameter-free statistic function to help the model capture key data statistics. Like an exponential family model, the hybrid model aims to align the distribution statistics with data statistics during model training, even when it only approximately maximizes the data likelihood. This property enables us to impose constraints on the hybrid model. Our empirical study validates the hybrid model's ability to match statistics. Furthermore, experimental results show that data fitting and generation improve when suitable informative statistics are incorporated into the hybrid model.
Similar Papers
Energy Matching: Unifying Flow Matching and Energy-Based Models for Generative Modeling
Machine Learning (CS)
Makes AI create better pictures with more control.
Exploring bidirectional bounds for minimax-training of Energy-based models
Machine Learning (CS)
Makes AI learn better and create clearer pictures.
Learning Energy-Based Generative Models via Potential Flow: A Variational Principle Approach to Probability Density Homotopy Matching
Machine Learning (CS)
Creates better AI art and understanding.