Random Signals - Part 2 - Updated
Random Signals - Part 2 - Updated
Axioms of Probability
Example 3.1.
(a) Flipping a coin: The sample space is Ω = {Head, Tail}.
(b) Rolling a die: The sample space is Ω = {1, 2, 3, 4, 5, 6}.
(c) Flipping two coins: The sample space is Ω = {HH, HT, T H, T T }
Probability law
Event B
Experiment
Event A
Sample Space
Events
10
Figure 3.2: An illustration of sample space, outcome, and events
Example 3.2. The rolling of a die forms a common experiment. The set
of prime numbers less than or equal to six, namely {2, 3, 5}, is one of many
possible events. The actual number observed after rolling the die is the outcome
of the experiment.
Note that the sample space should be (a) collectively exhaustive: it should
include all the possible outcomes and (b) mutually exclusive: if one outcome
happens the others may not.
(Non-negativity) Pr(A) ≥ 0
(Additivity) If A and B are two disjoint events, then the probability
of their union is
Pr(A ∪ B) = Pr(A) + Pr(B) (3.1)
Note: If events A and B are not disjoint, then the probability of their union
is given by
Pr(A ∪ B) = Pr(A) + Pr(B) − Pr(A ∩ B) (3.2)
11
Consider an experiment involving a single coin toss with the sample space
Ω = {H, T }. If the coin is fair, then P(A) = P(B). The additivity axiom
implies that
P(Ω) = P(H ∪ T ) = P(H) + P(T ) = 1
Hence,
P(H) = P(T ) = 0.5
Using additivity,
number of elements of A
P(A) = (3.3)
n
Ω = {(3, 1), (3, 2), (3, 3), (3, 4), (3, 5), (3, 6)}
12
with the (conditional) probability of each outcome being 1/6. Hence, the
probability of having a sum of 8 given that the first die is 3 equals 1/6.
If we let A and B denote the event that the sum is 8 and the event that
the first die is 3, respectively, then the probability obtained above is called the
conditional probability that A occurs given that B has occurred and is defined
as
Conditional Probability
If Pr(B) > 0
Pr(A, B)
Pr(A|B) = (3.4)
Pr(B)
Example 3.3. We toss a fair coin three successive times. What is the condi-
tional probability Pr{A|B} when A and B are the events
A = {more heads than tails come up} , B = {1st toss is head}
Solution: The sample space is given by
3
Pr{A, B} =
8
Pr{A, B} 3/8 3
Pr{A|B} = = =
Pr{B} 4/8 4
13
Solution: This experiment can be represented by a sequential representation.
Let A and B be the events
14
3.4 Total Probability Theorem and Bayes’ Rule
Total Probability Theorem
If A1 , A2 , ..., An are disjoint events that from a partition of the sample
space, then
This theorem is illustrated in Fig. 3.4. Initially, the sample space is partitioned
into a number of disjoint events Ai ’s. Then the probability that B occurs is a
weighted average of its conditional probability under each event.
Example 3.5. The probability of winning for a football team is 0.3 against
half of the teams (call them type 1), 0.4 against a quarter of the teams (type
2), and 0.5 against the remaining teams (type 3). What is the probability of
winning against a random team?
Solution: Let Ai be the event of playing against a type i team. Then,
Pr(A1 ) = 0.5, Pr(A2 ) = 0.25, Pr(A3 ) = 0.25
and let B be the event of winning. Hence,
Pr(B|A1 ) = 0.3, Pr(B|A2 ) = 0.4, Pr(B|A3 ) = 0.5,
Using total probability theorem, the probability of winning is
15
The total probability theorem is often used in conjunction with the fol-
lowing theorem, which relates conditional probabilities of the form Pr(A|B)
with conditional probabilities of the form Pr(B|A), in which the order of the
conditioning is reversed.
Bayes’ Rule
Let A1 , A2 , ..., An are disjoint events that from a partition of the sample
space. Then
Pr(Ai ) Pr(B|Ai )
Pr(Ai |B) = (3.7)
Pr(B)
Combining this with the total probability theorem results in
Pr(Ai ) Pr(B|Ai )
= (3.8)
Pr(A1 ) Pr(B|A1 ) + ... + Pr(An ) Pr(B|An )
Bayes’ rule is often used for inference. There are a number of ”causes” that
may result in a certain ”effect”. We observe the effect and wish to infer the
cause. The events A1 , A2 , ..., An are associated with the causes and the event
B represents the effect. Given that the effect B has been observed, we wish
to evaluate the probability that the cause Ai is present, i.e. Pr(Ai |B). We
refer to Pr(Ai |B) as the posterior probability of event Ai given the prior
probability Pr(Ai ). The cause-effect versus inference in illustrated in Fig.
3.5.
Example 3.6. In the radar problem of example 3.4, what is the probability
that an aircraft is present when an alarm is generated?
Solution: Solution: Let A and B be the events
Pr(A) Pr(B|A)
Pr(A|B) = (3.9)
Pr(B)
Pr(A) Pr(B|A)
= (3.10)
Pr(A) Pr(B|A) + Pr(AC ) Pr(B|AC )
0.05 × 0.99
= = 0.342 (3.11)
0.5 × 0.99 + 0.95 × 0.1
16
cause-effect
inference
Figure 3.5: Cause-effect and inference
Example 3.7. In the football problem of example 3.5, if the team wins what
is the probability that it played against a type 1 team?
Solution: Ai is the event of playing against a type i team,
Pr(A1 ) = 0.5, Pr(A2 ) = 0.25, Pr(A3 ) = 0.25
and B is the event of winning.
Pr(B|A1 ) = 0.3, Pr(B|A2 ) = 0.4, Pr(B|A3 ) = 0.5,
Using Bayes’ rule, the desired probability Pr(A1 |B) is given by
Pr(A1 ) Pr(B|A1 )
Pr(A1 |B) =
Pr(A1 ) Pr(B|A1 ) + Pr(A2 ) Pr(B|A2 ) + Pr(A3 ) Pr(B|A3 )
0.5 × 0.3
= = 0.4
0.5 × 0.3 + 0.25 × 0.4 + 0.25 × 0.5
3.5 Independence
The conditional probability Pr(A|B) captures the partial information that
event B provides about event A. An important special case arises when the oc-
currence of B provides no such information and does not alter the probability
that A has occurred, i.e.,
Pr(A|B) = Pr(A)
When the above equality holds, we say that A is independent of B. Given that
Pr(A|B) = Pr(A, B)/ Pr(B), this is equivalent to
17
A common first thought is that two events are independent if they are
disjoint, but in fact the opposite is true: two disjoint events A and B with
Pr(A) > 0 and Pr(B) > 0 are never independent, since their intersection
A ∩ B is empty and has probability 0. For example, an event A and its
complement AC are not independent (unless Pr(A) = 0 or Pr(A) = 1) , since
knowledge that A has occurred provides precise information about whether AC
has occurred.
(b) We have
1
Pr(A, B) = Pr({1, 4}) =
16
and
1
Pr(A) =
4
Pr(B) = Pr{(1, 4), (2, 3), (3, 2), (4, 1)} = 4/16
Hence,
Pr(Ai , Bj ) = Pr(Ai ) Pr(Bj )
18