Geometric Convergence Analysis of Variational Inference via Bregman Divergences
By: Sushil Bohara, Amedeo Roberto Esposito
Potential Business Impact:
Helps computers learn better by understanding math.
Variational Inference (VI) provides a scalable framework for Bayesian inference by optimizing the Evidence Lower Bound (ELBO), but convergence analysis remains challenging due to the objective's non-convexity and non-smoothness in Euclidean space. We establish a novel theoretical framework for analyzing VI convergence by exploiting the exponential family structure of distributions. We express negative ELBO as a Bregman divergence with respect to the log-partition function, enabling a geometric analysis of the optimization landscape. We show that this Bregman representation admits a weak monotonicity property that, while weaker than convexity, provides sufficient structure for rigorous convergence analysis. By deriving bounds on the objective function along rays in parameter space, we establish properties governed by the spectral characteristics of the Fisher information matrix. Under this geometric framework, we prove non-asymptotic convergence rates for gradient descent algorithms with both constant and diminishing step sizes.
Similar Papers
A Generalized Bias-Variance Decomposition for Bregman Divergences
Machine Learning (CS)
Makes computer learning more accurate for certain tasks.
Rates of Convergence of Generalised Variational Inference Posteriors under Prior Misspecification
Statistics Theory
Makes AI learn better even with wrong starting guesses.
Maxitive Donsker-Varadhan Formulation for Possibilistic Variational Inference
Machine Learning (Stat)
Lets computers learn better with less information.