Kullback-Leibler Divergence (KL Divergence), also known as relative entropy, is a statistical measure that quantifies the information loss or 'surprise' incurred when using an approximate probability distribution Q to represent a true distribution P. Formally, for discrete distributions, it is defined as D_KL(P || Q) = Σ P(x) log(P(x)/Q(x)). It is non-negative and zero only when P and Q are identical, but it is not a true distance metric as it is not symmetric (D_KL(P || Q) ≠ D_KL(Q || P)). This asymmetry makes it directional, measuring the inefficiency of assuming Q when P is true.
