Example: A Binary System Produces Marks With Probability of 0.7 and Spaces With Probability 0.3, 2/7 of The Marks Are Received in Error and 1/3 of The Spaces. Find The Information Trans-Fer
Example: A Binary System Produces Marks With Probability of 0.7 and Spaces With Probability 0.3, 2/7 of The Marks Are Received in Error and 1/3 of The Spaces. Find The Information Trans-Fer
1
p(y = M ) = p(M, M ) + p(S, M )
= 0.5 + 0.1 = 0.6
p(y = S) = p(S, S) + p(M, S)
= 0.2 + 0.2 = 0.4
Use
p(xi, yj )
I(xi; yj ) = log2
p(xi)p(yi)
to calculate mutual information for X = {M, S},
Y = {M, S}.
0.2 ) = −0.4854
I(X = M, Y = S) = log2( 0.7×0.4
0.5 ) = 0.2515
I(X = M, Y = M ) = log2( 0.7×0.6
0.2 ) = 0.7370
I(X = S, Y = S) = log2( 0.3×0.4
0.1 ) = −0.8480
I(X = S, Y = M ) = log2( 0.3×0.6
2
Use
XX
I(X; Y ) = p(xi, yj )I(xi; yj )
X Y
to calculate average mutual information
p(y|x)p(x)
To calculate p(x|y), p(x|y) = p(y)
was ap-
plied.
3
•Equivocation
so that
4
The information sent by the observer is easily
evaluated as −[p(0) log2 p(0) + p(1) log2 p(1)]
applied to the binary string generated by the
observer. The probability of 0 is just the chan-
nel error probability.
5
The general expression of the information trans-
fer using equivocation (information loss) is there-
fore
I(X; Y ) = H(X) − H(X|Y ) = H(Y ) − H(Y |X)
where
XX
H(Y |X) = − p(xi, yj ) log2 p(yj |xi)
X Y
XX
H(X|Y ) = − p(xi, yj ) log2 p(xi|yj )
X Y
H(X) = − X p(xi) log2 p(xi) is the source en-
P
tropy.
Venn diagram of the formulae
6
A summary of basic formulae
7
•Channel capacity
C = max I(X, Y )
For a Binary Symmetric Channels (BSC) (This
is characterized by a single value p of binary er-
ror probability, that is the probabilities of errors
in 0 and 1 is the same. )
I(X; Y )
= H(X) − H(X|Y )
= H(Y ) − H(Y |X)
XX
= H(Y ) + p(xi, yj ) log2 p(yj |xi)
X Y
X X
= H(Y ) + p(xi)p(yj |xi) log2 p(yj |xi)
X Y
8
Continue ...
I(X; Y )
XX
= H(Y ) + p(xi )p(yj |xi ) log2 p(yj |xi )
X Y
X X
= H(Y ) + p(xi ) p(yj |xi ) log2 p(yj |xi )
X Y
X
= H(Y ) + p(xi )[p log2 p + (1 − p) log2 (1 − p)]
X
X
= H(Y ) + [p log2 p + (1 − p) log2 (1 − p)] p(xi )
X
= H(Y ) + [p log2 p + (1 − p) log2 (1 − p)]
= H(Y ) − H(p)
C = 1 − H(p)
9
For example, if p = 0.125, then
H(p) = −[p log2 p + (1 − p) log2(1 − p)]
= 0.54 bits
0.7
0.6
0.5
I(xy)
0.4
0.3
0.2
0.1
−0.1
0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
p(0)
0.8
0.6
C
0.4
0.2
0
0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
error probability p
11