An Inertial Langevin Algorithm
By: Alexander Falk , Andreas Habring , Christoph Griesbacher and more
Potential Business Impact:
Makes computer models learn faster and better.
We present a novel method for drawing samples from Gibbs distributions with densities of the form $\pi(x) \propto \exp(-U(x))$. The method accelerates the unadjusted Langevin algorithm by introducing an inertia term similar to Polyak's heavy ball method, together with a corresponding noise rescaling. Interpreting the scheme as a discretization of \emph{kinetic} Langevin dynamics, we prove ergodicity (in continuous and discrete time) for twice continuously differentiable, strongly convex, and $L$-smooth potentials and bound the bias of the discretization to the target in Wasserstein-2 distance. In particular, the presented proofs allow for smaller friction parameters in the kinetic Langevin diffusion compared to existing literature. Moreover, we show the close ties of the proposed method to the over-relaxed Gibbs sampler. The scheme is tested in an extensive set of numerical experiments covering simple toy examples, total variation image denoising, and the complex task of maximum likelihood learning of an energy-based model for molecular structure generation. The experimental results confirm the acceleration provided by the proposed scheme even beyond the strongly convex and $L$-smooth setting.
Similar Papers
Anchored Langevin Algorithms
Machine Learning (Stat)
Helps computers learn from tricky, uneven data.
Contractive kinetic Langevin samplers beyond global Lipschitz continuity
Probability
Makes computer models learn faster and more accurately.
Operator-Level Quantum Acceleration of Non-Logconcave Sampling
Quantum Physics
Quantum computers solve tough problems faster.