Particle Hamiltonian Monte Carlo
By: Alaa Amri, Víctor Elvira, Amy L. Wilson
Potential Business Impact:
Helps computers understand complex patterns better.
In Bayesian inference, Hamiltonian Monte Carlo (HMC) is a popular Markov Chain Monte Carlo (MCMC) algorithm known for its efficiency in sampling from complex probability distributions. However, its application to models with latent variables, such as state-space models, poses significant challenges. These challenges arise from the need to compute gradients of the log-posterior of the latent variables, and the likelihood may be intractable due to the complexity of the underlying model. In this paper, we propose Particle Hamiltonian Monte Carlo (PHMC), an algorithm specifically designed for state-space models. PHMC leverages Sequential Monte Carlo (SMC) methods to estimate the marginal likelihood, infer latent variables (as in particle Metropolis-Hastings), and compute gradients of the log-posterior of model parameters. Importantly, PHMC avoids the need to calculate gradients of the log-posterior for latent variables, which addresses a major limitation of traditional HMC approaches. We assess the performance of Particle HMC on both simulated datasets and a real-world dataset involving crowdsourced cycling activities data. The results demonstrate that Particle HMC outperforms particle marginal Metropolis-Hastings with a Gaussian random walk, particularly in scenarios involving a large number of parameters.
Similar Papers
Proximal Hamiltonian Monte Carlo
Computation
Improves computer image cleaning and signal finding.
Particle Gibbs without the Gibbs bit
Computation
Improves how computers guess hidden patterns.
Bouncy particle sampler with infinite exchanging parallel tempering
Machine Learning (CS)
Makes computer predictions more accurate with faster sampling.