Lec2 - Data Compression PDF
Lec2 - Data Compression PDF
𝐻 𝑥 = 𝐸 𝐼 𝑋𝑖
𝑀 𝑏𝑖𝑡𝑠
= − 𝑃𝑖 𝑙𝑜𝑔2 𝑃𝑖
𝑖=1 𝑠𝑦𝑚𝑏𝑜𝑙
And E{} is the expected value of X
Example: Entropy of a binary
Source
• Assume that a binary source generate ones and zeros
with probability of 0,1 and 0.9 repectively
• 𝐻 𝑥 = − σ2𝑖=1 𝑃𝑖 𝑙𝑜𝑔2 (𝑃𝑖 ) = 0.1𝑙𝑜𝑔2 0.1 + 0.9𝑙𝑜𝑔2 0.9
• =0.47 bis/symbol
• The result obtained because information depends on
the probability , small probability contains information
higher than large probability.
• However , the upper bound is 𝑙𝑜𝑔2 M and this always
sttitafy the equality below
• 0 ≤ 𝐻 𝑥 ≤ 𝑙𝑜𝑔2 (𝑀) where M is number of Symbols
Discrete Sources (continued
• Memoryless
A discrete source is said to be memoryless if the symbol emitted by the source are statistically independent .
Example if two symbol is chosen at a time, the joint probability of the two elements is simlpy the product of
their probabilities
This is also called source Entropy and E(X) is the expected value of X.
With Memory
The discrete is said to have memory if the source elements composing the sequence are not independents
𝐻 𝑥 ≤ 𝑙𝑖𝑚𝑀→∞ 𝐻𝑀 𝑋 𝑎𝑛𝑑 𝐻𝑀 (𝑋)𝑚𝑒𝑚𝑜𝑟𝑦 < 𝐻𝑀 (𝑋)𝑛𝑜𝑚𝑒𝑚𝑜𝑟𝑦
In conclusion the average entropy per symbol of the M-tuple from a` source with memory decreases
as the length M increases.
It is more efficient to encode symols from a source with memory in groups of several symbols rather
than to encode them one at a time.
Example 2