From Encyclopedia of Mathematics
Revision as of 18:15, 19 October 2014 by Richard Pinch (talk | contribs) (Category:Information and communication, circuits)
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Jump to: navigation, search

A measure of the possible increase in the transmission rate of information by using a statistical dependence between the components of the information processed at the source of information. The redundancy of a stationary source of information in discrete time processing the information $\xi = ( \ldots, \xi_{-1}, \xi_0, \xi_1, \ldots )$ generated by a stationary stochastic process $$ \xi_k\,,\ \ \ k = \ldots, -1,0,1, \ldots $$ where$\xi_k$ takes values in some finite set $X$ with $N$ elements, is defined to be $$ 1 - \frac{\bar H(U)}{H_{\mathrm{max}}} $$ where $\bar H(U)$ is the rate of generation of information by the given source $U$ (see Information, rate of generation of) and $H_{\mathrm{max}} = \log N$ is the maximum possible speed of generation of information by a source in discrete time whose components take $N$ different values.

For references, see at Communication channel.

How to Cite This Entry:
Redundancy. Encyclopedia of Mathematics. URL:
This article was adapted from an original article by R.L. DobrushinV.V. Prelov (originator), which appeared in Encyclopedia of Mathematics - ISBN 1402006098. See original article