Generative modelling with jump-diffusions
By: Adrian Baule
Potential Business Impact:
Makes AI create more realistic pictures and sounds.
Score-based diffusion models generate samples from an unknown target distribution using a time-reversed diffusion process. While such models represent state-of-the-art approaches in industrial applications such as artificial image generation, it has recently been noted that their performance can be further improved by considering injection noise with heavy tailed characteristics. Here, I present a generalization of generative diffusion processes to a wide class of non-Gaussian noise processes. I consider forward processes driven by standard Gaussian noise with super-imposed Poisson jumps representing a finite activity Levy process. The generative process is shown to be governed by a generalized score function that depends on the jump amplitude distribution. Both probability flow ODE and SDE formulations are derived using basic technical effort, and are implemented for jump amplitudes drawn from a multivariate Laplace distribution. Remarkably, for the problem of capturing a heavy-tailed target distribution, the jump-diffusion Laplace model outperforms models driven by alpha-stable noise despite not containing any heavy-tailed characteristics. The framework can be readily applied to other jump statistics that could further improve on the performance of standard diffusion models.
Similar Papers
An effective potential for generative modelling with active matter
Statistical Mechanics
Makes AI create realistic pictures from simple ideas.
An effective potential for generative modelling with active matter
Statistical Mechanics
Creates new data using tiny moving things.
Score-based constrained generative modeling via Langevin diffusions with boundary conditions
Machine Learning (Stat)
Makes AI create images that follow rules.