## Encyclopedia > Data compression entropy

Article Content

# Information entropy

Redirected from Data compression/entropy

Entropy is a concept in thermodynamics (see entropy) and information theory. The two concepts do actually have something in common, although it takes a thorough understanding of both fields for this to become apparent.

Claude E. Shannon defined a measure of entropy (H = - Σ pi log pi) that, when applied to an information source, could determine the capacity of the channel required to transmit the source as encoded binary digits. Shannon's measure of entropy came to be taken as a measure of the information contained in a message, as opposed to the portion of the message that is strictly determined (hence predictable) by inherent structures, like for instance redundancy in the structure of languages or the statistical properties of a language relating to the frequencies of occurrence of different letter or word pairs, triplets etc. See Markov chains.

Entropy as defined by Shannon is closely related to entropy as defined by physicists. Boltzmann and Gibbs did considerable work on statistical thermodynamics. This work was the inspiration for adopting the term entropy in information theory. There are deep relationships between entropy in the thermodynamic and informational senses. For instance, Maxwell's demon needs information to reverse thermodynamic entropy and getting that information exactly balances out the thermodynamic gain that the demon would otherwise achieve.

In information theory, entropy is conceptually the actual amount of (information theoretic) infomation in a piece of data. Entirely random byte data has an entropy of about 8, since you never know what the next character will be. A long string of A's has an entropy of 0, since you know that the next character will always be an 'A'. The entropy of English text is about 1.5 bits per character (Try compressing it with the PPM compression algorithm[?]!) The entropy rate of a data source means the average number of bits per symbol needed to encode it.

1. Many of the bits in the data may not be conveying any infomation. For instance it is often the case that data structures store information redundantly, or have sections that are always the same regardless of the information in the data structure.
2. The amount of entropy is not always an integer number of bits.

Entropy is effectively the lowest compression possible, which can be realised in theory by the use of the typical set or in practise using Huffman, Lempel-Ziv or Arithmetic coding. The definition of entropy is based on the Markov model of text. For an order-0 source (each character is selected independent of the last characters), the entropy is:

$H(S) = - \sum p_i \log_2 p_i$

Where $p_i$ is the probability of $i$. For a higher-order Markov source (one in which probabilities are dependent on the preceding characters), the entropy is:

$H(S) = \sum_i \sum_j p_i \ p_i (j) \log_2 p_i (j) = \sum_i \sum_j p_{ij} \log_2 p_i(j)$

Where $i$ is a state (certain preceding characters) and $p_i(j)$ is the probability of $j$ given $i$ as the previous character (s).

All Wikipedia text is available under the terms of the GNU Free Documentation License

Search Encyclopedia
 Search over one million articles, find something about almost anything!

Featured Article
 22nd century ... from the century known as the twenty-one hundreds, which comprises the years 2100-2199. Decades and ...