Redundancy (information theory)
   HOME

TheInfoList



OR:

In information theory, redundancy measures the fractional difference between the
entropy Entropy is a scientific concept, as well as a measurable physical property, that is most commonly associated with a state of disorder, randomness, or uncertainty. The term and the concept are used in diverse fields, from classical thermodynam ...
of an ensemble , and its maximum possible value \log(, \mathcal_X, ). Informally, it is the amount of wasted "space" used to transmit certain data.
Data compression In information theory, data compression, source coding, or bit-rate reduction is the process of encoding information using fewer bits than the original representation. Any particular compression is either lossy or lossless. Lossless compressio ...
is a way to reduce or eliminate unwanted redundancy, while forward error correction is a way of adding desired redundancy for purposes of error detection and correction when communicating over a noisy channel of limited capacity.


Quantitative definition

In describing the redundancy of raw data, the rate of a source of information is the average
entropy Entropy is a scientific concept, as well as a measurable physical property, that is most commonly associated with a state of disorder, randomness, or uncertainty. The term and the concept are used in diverse fields, from classical thermodynam ...
per symbol. For memoryless sources, this is merely the entropy of each symbol, while, in the most general case of a stochastic process, it is :r = \lim_ \frac H(M_1, M_2, \dots M_n), in the limit, as ''n'' goes to infinity, of the
joint entropy In information theory, joint entropy is a measure of the uncertainty associated with a set of variables. Definition The joint Shannon entropy (in bits) of two discrete random variables X and Y with images \mathcal X and \mathcal Y is defined ...
of the first ''n'' symbols divided by ''n''. It is common in information theory to speak of the "rate" or "
entropy Entropy is a scientific concept, as well as a measurable physical property, that is most commonly associated with a state of disorder, randomness, or uncertainty. The term and the concept are used in diverse fields, from classical thermodynam ...
" of a language. This is appropriate, for example, when the source of information is English prose. The rate of a memoryless source is simply H(M), since by definition there is no interdependence of the successive messages of a memoryless source. The absolute rate of a language or source is simply :R = \log , \mathbb M, ,\, the
logarithm In mathematics, the logarithm is the inverse function to exponentiation. That means the logarithm of a number  to the base  is the exponent to which must be raised, to produce . For example, since , the ''logarithm base'' 10 of ...
of the cardinality of the message space, or alphabet. (This formula is sometimes called the
Hartley function The Hartley function is a measure of uncertainty, introduced by Ralph Hartley in 1928. If a sample from a finite set ''A'' uniformly at random is picked, the information revealed after the outcome is known is given by the Hartley function : H_0(A) ...
.) This is the maximum possible rate of information that can be transmitted with that alphabet. (The logarithm should be taken to a base appropriate for the unit of measurement in use.) The absolute rate is equal to the actual rate if the source is memoryless and has a uniform distribution. The absolute redundancy can then be defined as : D = R - r ,\, the difference between the absolute rate and the rate. The quantity \frac D R is called the relative redundancy and gives the maximum possible
data compression ratio Data compression ratio, also known as compression power, is a measurement of the relative reduction in size of data representation produced by a data compression algorithm. It is typically expressed as the division of uncompressed size by compresse ...
, when expressed as the percentage by which a file size can be decreased. (When expressed as a ratio of original file size to compressed file size, the quantity R : r gives the maximum compression ratio that can be achieved.) Complementary to the concept of relative redundancy is efficiency, defined as \frac r R , so that \frac r R + \frac D R = 1. A memoryless source with a uniform distribution has zero redundancy (and thus 100% efficiency), and cannot be compressed.


Other notions

A measure of ''redundancy'' between two variables is the
mutual information In probability theory and information theory, the mutual information (MI) of two random variables is a measure of the mutual dependence between the two variables. More specifically, it quantifies the " amount of information" (in units such ...
or a normalized variant. A measure of redundancy among many variables is given by the total correlation. Redundancy of compressed data refers to the difference between the expected compressed data length of n messages L(M^n) \,\! (or expected data rate L(M^n)/n \,\!) and the entropy nr \,\! (or entropy rate r \,\!). (Here we assume the data is ergodic and stationary, e.g., a memoryless source.) Although the rate difference L(M^n)/n-r \,\! can be arbitrarily small as n \,\! increased, the actual difference L(M^n)-nr \,\!, cannot, although it can be theoretically upper-bounded by 1 in the case of finite-entropy memoryless sources. Redundancy in an information-theoretic contexts can also refer to the information that is redundant between two mutual informations. For example, given three variables X_1, X_2, and Y, it is known that the joint mutual information can be less than the sum of the marginal mutual informations: I(X_1,X_2 ; Y) < I(X_1;Y) + I(X_2;Y). In this case, at least some of the information about Y disclosed by X_1 or X_2 is the same. This formulation of redundancy is complementary to the notion of synergy, which occurs when the joint mutual information is greater than the sum of the marginals, indicating the presence of information that is only disclosed by the joint state and not any simpler collection of sources.


See also

*
Minimum redundancy coding In information theory, an entropy coding (or entropy encoding) is any lossless data compression method that attempts to approach the lower bound declared by Shannon's source coding theorem, which states that any lossless data compression method m ...
**
Huffman encoding In computer science and information theory, a Huffman code is a particular type of optimal prefix code that is commonly used for lossless data compression. The process of finding or using such a code proceeds by means of Huffman coding, an algor ...
*
Data compression In information theory, data compression, source coding, or bit-rate reduction is the process of encoding information using fewer bits than the original representation. Any particular compression is either lossy or lossless. Lossless compressio ...
*
Hartley function The Hartley function is a measure of uncertainty, introduced by Ralph Hartley in 1928. If a sample from a finite set ''A'' uniformly at random is picked, the information revealed after the outcome is known is given by the Hartley function : H_0(A) ...
*
Negentropy In information theory and statistics, negentropy is used as a measure of distance to normality. The concept and phrase "negative entropy" was introduced by Erwin Schrödinger in his 1944 popular-science book ''What is Life?'' Later, Léon Brillo ...
*
Source coding theorem In information theory, Shannon's source coding theorem (or noiseless coding theorem) establishes the limits to possible data compression, and the operational meaning of the Shannon entropy. Named after Claude Shannon, the source coding theorem ...
* Overcompleteness


References

* * * {{Compression Methods Information theory