# Kullback-Leibler information

*Kullback–Leibler quantity of information, Kullback–Leibler information quantity, directed divergence*

For discrete distributions (cf. Discrete distribution) given by probability vectors , , the Kullback–Leibler (quantity of) information of with respect to is:

where is the natural logarithm (cf. also Logarithm of a number).

More generally, one has:

for probability distributions and with densities and (cf. Density of a probability distribution).

The negative of is the conditional entropy (or relative entropy) of with respect to ; see Entropy.

Various notions of (asymmetric and symmetric) information distances are based on the Kullback–Leibler information.

The quantity is also called the informational divergence (see Huffman code).

See also Information distance; Kullback–Leibler-type distance measures.

#### References

[a1] | S. Kullback, "Information theory and statistics" , Wiley (1959) |

[a2] | S. Kullback, R.A. Leibler, "On information and sufficiency" Ann. Math. Stat. , 22 (1951) pp. 79–86 |

[a3] | J. Sakamoto, M. Ishiguro, G. Kitagawa, "Akaike information criterion statistics" , Reidel (1986) |

**How to Cite This Entry:**

Kullback-Leibler information.

*Encyclopedia of Mathematics.*URL: http://encyclopediaofmath.org/index.php?title=Kullback-Leibler_information&oldid=22681