Comparing Misspecified Models with Big Data: A Variational Bayesian Perspective
By: Yong Li , Sushanta K. Mallick , Tao Zeng and more
Potential Business Impact:
Helps computers pick the right data faster.
Optimal data detection in massive multiple-input multiple-output (MIMO) systems often requires prohibitively high computational complexity. A variety of detection algorithms have been proposed in the literature, offering different trade-offs between complexity and detection performance. In recent years, Variational Bayes (VB) has emerged as a widely used method for addressing statistical inference in the context of massive data. This study focuses on misspecified models and examines the risk functions associated with predictive distributions derived from variational posterior distributions. These risk functions, defined as the expectation of the Kullback-Leibler (KL) divergence between the true data-generating density and the variational predictive distributions, provide a framework for assessing predictive performance. We propose two novel information criteria for predictive model comparison based on these risk functions. Under certain regularity conditions, we demonstrate that the proposed information criteria are asymptotically unbiased estimators of their respective risk functions. Through comprehensive numerical simulations and empirical applications in economics and finance, we demonstrate the effectiveness of these information criteria in comparing misspecified models in the context of massive data.
Similar Papers
Detecting Model Misspecification in Bayesian Inverse Problems via Variational Gradient Descent
Methodology
Finds when computer models are wrong.
Robust and Scalable Variational Bayes
Machine Learning (Stat)
Cleans messy data for better computer learning.
Variational Inference with Mixtures of Isotropic Gaussians
Machine Learning (Stat)
Finds better computer guesses for complex problems.