Kullback-Leibler divergence

from Wikipedia, the free encyclopedia

The terms Kullback-Leibler divergence ( KL divergence for short ), Kullback-Leibler entropy , Kullback-Leibler information , Kullback-Leibler distance (according to Solomon Kullback and Richard Leibler ) or information gain denote a measure of the difference between two probability distributions . Typically, one of the distributions represents empirical observations or a precise probability distribution, while the other represents a model or an approximation.

Note : The KL divergence is also called relative entropy , whereby the term relative entropy is occasionally also used for transinformation .

Formally, the KL divergence for the probability functions and discrete values ​​can be determined as follows:

If the distributions and for continuous values ​​are represented by the probability density functions and , an integral is calculated:

The Kullback-Leibler divergence is from information theory to view how much space is wasted per character on average when one on based coding is applied to a source of information that the actual distribution follows. There is thus a connection to the channel capacity . Mathematically, this is compatible with the statement that the KL is divergence and equality only applies if P and Q are identical.

The specific choice of the base of the logarithm in the calculation depends on the information unit to be used for the calculation. In practice, the KL divergence is often given in bits or Shannon and base 2 is used for this, nit (base ) and ban (base 10) are also used more rarely .

Instead of the Kullback-Leibler divergence, the cross entropy is often used. This provides qualitatively comparable values, but can be estimated without precise knowledge of . In practical applications this is advantageous because the actual background distribution of the observation data is mostly unknown.

Although the Kullback-Leibler divergence is sometimes also referred to as the Kullback-Leibler distance, it does not meet a fundamental requirement for distance measures : It is not symmetrical, so it applies in general . Alternatively, to establish symmetry, the sum of the two divergences can be used, which is obviously symmetrical:

Multivariate normal distributions

For two multi-dimensional normal distributions (with dimension ), with mean values and (non-singular) covariance matrices , the Kullback-Leibler divergence is given by:

example

An example of the application of the Kullback-Leibler distance to assess the similarity of two probability distributions is the comparison of two sets of grooves on a honed surface. For the grooves, a parametric model is first set up for the depth, spacing and width of the grooves. The parameters for the samples to be compared are then estimated. Now the Kullback-Leibler distance between the two distributions can be calculated. Their size indicates the similarity of the distributions of the groups of grooves; if it is 0, both samples come from the same distribution.

supporting documents

Individual evidence

  1. Duchi J., “ Derivations for Linear Algebra and Optimization, page 13 ”.
  2. Doris Krahe, Jürgen Beyerer: Parametric method to quantify the balance of groove sets of honed cylinder bores . In: Architectures, Networks, and Intelligent Systems for Manufacturing Integration . doi : 10.1117 / 12.294431 ( spie.org [accessed August 2, 2016]).