Backpropagation-Free Metropolis-Adjusted Langevin Algorithm
By: Adam D. Cobb, Susmit Jha
Potential Business Impact:
Teaches computers without needing to show every step.
Recent work on backpropagation-free learning has shown that it is possible to use forward-mode automatic differentiation (AD) to perform optimization on differentiable models. Forward-mode AD requires sampling a tangent vector for each forward pass of a model. The result is the model evaluation with the directional derivative along the tangent. In this paper, we illustrate how the sampling of this tangent vector can be incorporated into the proposal mechanism for the Metropolis-Adjusted Langevin Algorithm (MALA). As such, we are the first to introduce a backpropagation-free gradient-based Markov chain Monte Carlo (MCMC) algorithm. We also extend to a novel backpropagation-free position-specific preconditioned forward-mode MALA that leverages Hessian information. Overall, we propose four new algorithms: Forward MALA; Line Forward MALA; Pre-conditioned Forward MALA, and Pre-conditioned Line Forward MALA. We highlight the reduced computational cost of the forward-mode samplers and show that forward-mode is competitive with the original MALA, while even outperforming it depending on the probabilistic model. We include Bayesian inference results on a range of probabilistic models, including hierarchical distributions and Bayesian neural networks.
Similar Papers
Metropolis-adjusted Subdifferential Langevin Algorithm
Methodology
Lets computers learn from messy, uneven information.
Solving Bayesian inverse problems via Fisher adaptive Metropolis adjusted Langevin algorithm
Numerical Analysis
Helps computers solve hard math problems faster.
A Langevin sampling algorithm inspired by the Adam optimizer
Computation
Helps computers learn faster and more accurately.