Variational Bayesian methods are a family of techniques for approximating intractable
integrals arising in
Bayesian inference and
machine learning. They are typically used in complex
statistical models consisting of observed variables (usually termed "data") as well as unknown
parameters and
latent variables, with various sorts of relationships among the three types of
random variables, as might be described by a
graphical model. As is typical in Bayesian inference, the parameters and latent variables are grouped together as "unobserved variables". Variational Bayesian methods are primarily used for two purposes:
- To provide an analytical approximation to the posterior probability of the unobserved variables, in order to do statistical inference over these variables.
- To derive a lower bound for the marginal likelihood (sometimes called the "evidence") of the observed data (i.e. the marginal probability of the data given the model, with marginalization performed over unobserved variables). This is typically used for performing model selection, the general idea being that a higher marginal likelihood for a given model indicates a better fit of the data by that model and hence a greater probability that the model in question was the one that generated the data. (See also the Bayes factor article.)