Easily Computed Marginal Likelihoods for Multivariate Mixture Models Using the THAMES Estimator
By: Martin Metodiev , Nicholas J. Irons , Marie Perrot-Dockès and more
Potential Business Impact:
Helps computers understand complex data better.
We present a new version of the truncated harmonic mean estimator (THAMES) for univariate or multivariate mixture models. The estimator computes the marginal likelihood from Markov chain Monte Carlo (MCMC) samples, is consistent, asymptotically normal and of finite variance. In addition, it is invariant to label switching, does not require posterior samples from hidden allocation vectors, and is easily approximated, even for an arbitrarily high number of components. Its computational efficiency is based on an asymptotically optimal ordering of the parameter space, which can in turn be used to provide useful visualisations. We test it in simulation settings where the true marginal likelihood is available analytically. It performs well against state-of-the-art competitors, even in multivariate settings with a high number of components. We demonstrate its utility for inference and model selection on univariate and multivariate data sets.
Similar Papers
Approximating evidence via bounded harmonic means
Computation
Improves computer guesses about which idea is best.
Near-Optimal Clustering in Mixture of Markov Chains
Machine Learning (Stat)
Groups similar paths by finding hidden patterns.
Uniform Mean Estimation for Heavy-Tailed Distributions via Median-of-Means
Machine Learning (Stat)
Finds averages in tricky data better.