Variational inference is a deterministic optimization technique used to approximate a complex, intractable posterior distribution in Bayesian statistics. Instead of directly computing the true posterior, VI introduces a simpler, parameterized family of distributions, known as the variational distribution or variational posterior, and optimizes its parameters to be as close as possible to the true posterior. This closeness is measured using the Kullback-Leibler (KL) divergence, a statistical measure of how one probability distribution diverges from another. The optimization objective is the Evidence Lower Bound (ELBO), a surrogate function that is maximized to minimize the KL divergence, thereby fitting the approximate distribution to the true one.
