0% found this document useful (0 votes)
84 views12 pages

Mathematical Expectation: Examples

The document provides examples and definitions for mathematical expectation, variance, and covariance. It defines: 1) Mathematical expectation as the average value of a random variable weighted by its probabilities. 2) Variance as a measure of how far a random variable's values are spread out from its expected value. 3) Covariance as a measure of how two random variables change together. It then proves properties of mathematical expectation, including that the expectation of a sum of random variables is equal to the sum of their individual expectations.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
84 views12 pages

Mathematical Expectation: Examples

The document provides examples and definitions for mathematical expectation, variance, and covariance. It defines: 1) Mathematical expectation as the average value of a random variable weighted by its probabilities. 2) Variance as a measure of how far a random variable's values are spread out from its expected value. 3) Covariance as a measure of how two random variables change together. It then proves properties of mathematical expectation, including that the expectation of a sum of random variables is equal to the sum of their individual expectations.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 12

Econ 326 Section 004

Notes on Mathematical Expectation, Variance, and Covariance


By Hiro Kasahara

Mathematical Expectation: Examples


• Consider the following game of chance. You pay 2 dollars and roll a fair die. Then you
receive a payment according to the following schedule. If the event A = {1, 2, 3} occurs, then
you will receive 1 dollar. If the event B = {4, 5} occurs, you receive 2 dollars. If the event
C = {6} occurs, then you will receive 6 dollars. What is the average profit you can make if
you participate this game?

If A occurs, then a profit will be 1 − 2 = −1 dollar, i.e., you will lose 1 dollar. If B occurs,
a profit will be 2 − 2 = 0. If C occurs, a profit will be 6 − 2 = 4 dollars. Therefore, we may
compute the average profit as follows:

average profit = (1/6+1/6+1/6)×(−1)+(1/6+1/6)×0+(1/6)×4 = (1/6)×(−3+0+4) = 1/6.

That is, you can expect to make 1/6 dollar on the average every time you play this game.
This is the mathematical expectation of the payment.

We can define a random variable X which represents a profit, where X takes a value of −1,
0, and 4 with probabilities 1/2, 1/3, and 1/6, respectively. Namely, P (X = −1) = 1/2,
P (X = 0) = 1/3, and P (X = 4) = 1/6. Then this mathematical expectation is written as
X
E(X) = xP (X = x) = (−1) × (1/2) + 0 × (1/3) + 4 × (1/6) = 1/6.
x∈{−1,0,4}

• Roll a die twice. Let X be the number of times 4 comes up. X takes three possible values
0, 1, or 2. X = 0 when the event {1, 2, 3, 5, 6} occurs for both cases so that P (X = 0) =
(5/6) × (5/6) = 25/36. X = 1 either when the event {1, 2, 3, 5, 6} occurs for the first die and
the event {4} occurs for the second die or when the event {4} occurs for the first die and the
event {1, 2, 3, 5, 6} occurs for the second die so that P (X = 1) = (5/6)×(1/6)+(1/6)×(5/6) =
10/36. Finally, X = 2 when the event {4} for both dies so that P (X = 2) = (1/6) × (1/6) =
1/36. Note that P (X = 0) + P (X = 1) + P (X = 2) = 1. Therefore, the mathematical
expectation of X is
X
E(X) = xP (X = x) = 0 × (25/36) + 1 × (10/36) + 2 × (1/36) = 1/3.
x=0,1,2

• Toss a coin 3 times. Let X be the number of heads. There are 8 possible outcomes:
{T T T, T T H, T HT, T HH, HT T, HT H, HHT, HHH}, where H indicates “Head” and T in-
dicates “Tail” X takes four possible values 0, 1, 2, and 3 with probabilities P (X = 0) = 1/8,
P (X = 1) = 3/8, P (X = 2) = 3/8, and P (X = 3) = 1/8. Therefore, the mathematical
expectation of X is
X
E(X) = xP (X = x) = 0×(1/8)+1×(3/8)+2×(3/8)+3×(1/8) = (0+3+6+3)/8 = 12/8 = 3/2.
x=0,1,2,3

1
Properties of Mathematical Expectation
Let X be a random variable and suppose that the mathematical expectation of X, E(X), exists.
1. If a is a constant, then
E(a) = a.
2. If b is a constant, then
E(bX) = bE(X).
3. If a and b are constants, then
E(a + bX) = a + bE(X). (1)

Proof: Let X be a discrete random variable, where possible values for X is {x1 , . . . , xn } with
probability mass function of X given by
pX
i = P (X = xi ) , i = 1, . . . n.
For the proof of 1, we have
n
X
E(a) = apX
i
i=1
= (apX X X
1 + ap2 + ... + apn )
= a × (pX X X
1 + p2 + ... + pn )
Xn
=a pX
i
i=1
=a
Pn X
where the last equality holds because i=1 pi = 1.
For the proof of 2, we have
n
X
E(bX) = bxi pX
i
i=1
= (bx1 pX X X
1 + bx2 p2 + .... + bxn pn )
= b × (x1 pX X X
1 + x2 p2 + .... + xn pn )
Xn
=b xi p X
i
i=1
= bE(X).
For the proof of 3, we have
n
X
E(a + bX) = (a + bxi )pX
i
i=1
= (a + bx1 )pX X X
1 + (a + bx2 )p2 + .... + (a + bxn )pn
= (apX X X X X X
1 + ap2 + ... + apn ) + (bx1 p1 + bx2 p2 + .... + bxn pn )
= a × (pX X X X X X
1 + p2 + ... + pn ) + b × (x1 p1 + x2 p2 + .... + xn pn )
Xn Xn
=a pX
i + b xi p X
i
i=1 i=1
= a + bE(X).

2
Variance and Covariance
Let X and Y be two discrete random variables. The set of possible values for X is {x1 , . . . , xn };
and the set of possible values for Y is {y1 , . . . , ym }. The joint probability function is given by

pX,Y
ij = P (X = xi , Y = yj ) , i = 1, . . . n; j = 1, . . . , m.

The marginal probability function of X is


m
pX,Y
X
pX
i = P (X = xi ) = ij , i = 1, . . . n,
j=1

and the marginal probability function of Y is


n
pX,Y
X
pYj = P (Y = yj ) = ij , j = 1, . . . m.
i=1

1.
E[X + Y ] = E[X] + E[Y ]. (2)

Proof:
n X
m
(xi + yj )pX,Y
X
E(X + Y ) = ij
i=1 j=1
n X m
(xi pX,Y + yj pX,Y
X
= ij ij )
i=1 j=1
n X
m n X
m
xi pX,Y yj pX,Y
X X
= ij + ij (3)
i=1 j=1 i=1 j=1
 
n m m n
!
pX,Y pX,Y
X X X X
= xi ·  ij
+ yj · ij (4)
i=1 j=1 j=1 i=1
Pm
because we can take xi out of j=1 because xi does not depend on j’s
n
X m
X
= xi · pX
i + yj · pYj
i=1 j=1
Pm X,Y Pn X,Y
because pX
i = j=1 pij and pYj = i=1 pij
= E(X) + E(Y )

X,Y X,Y X,Y Pn Pm X,Y


Equation (3): To understand ni=1 m
P P Pn Pm
j=1 (xi pij +yj pij ) = i=1 j=1 xi pij + i=1 j=1 yj pij ,
consider the case of n = m = 2. Then,
2 X
2
(xi pX,Y + yj pX,Y
X
ij ij )
i=1 j=1

= (x1 pX,Y X,Y X,Y X,Y X,Y X,Y X,Y X,Y


11 + y1 p11 ) + (x1 p12 + y2 p12 ) + (x2 p21 + y1 p21 ) + (x2 p22 + y2 p22 )
= (x1 pX,Y X,Y X,Y X,Y X,Y X,Y X,Y X,Y
11 + x1 p12 + x2 p21 + x2 p22 ) + (y1 p11 + y2 p12 + y1 p21 + y2 p22 )
2 X
2 2 X
2
xi pX,Y yj pX,Y
X X
= ij + ij .
i=1 j=1 i=1 j=1

3
Pn Pm X,Y Pn Pm X,Y
Equation (4): To understand i=1 j=1 xi pij = i=1 xi ·( j=1 pij ), consider the case
of n = m = 2. Then,
2 X
2
xi pX,Y = x1 pX,Y X,Y X,Y X,Y
X
ij 11 + x1 p12 + x2 p21 + x2 p22
i=1 j=1

= x1 (pX,Y X,Y X,Y X,Y


11 + p12 ) + x2 (p21 + p22 )
2
xi (pX,Y + pX,Y
X
= i1 i2 )
i=1
2 2
pX,Y
X X
= xi ( ij ).
i=1 j=1

X,Y
· ( 2i=1 pX,Y
P2 P2 P2 P
Similarly, we may show that i=1 j=1 yj pij = j=1 yj ij ).

2. If c is a constant, then Cov (X, c) = 0.

Proof: According to the definition of covariance,


Cov(X, c) = E[(X − E(X))(c − E(c))].
Since the expectation of a constant is itself, i.e., E(c) = c,
Cov(X, c) = E[(X − E(X))(c − c)]
= E[(X − E(X)) · 0]
= E[0]
Xn
= 0 × pX
i
i=1
Xn
= 0
i=1
= 0 + 0 + ... + 0
=0

3. Cov (X, X) = V ar (X) .

Proof: According to the definition of covariance, we can expand Cov(X, X) as follows:


Cov(X, X) = E[(X − E(X))(X − E(X))]
Xn n
X
= [xi − E(X)][xi − E(X)] · P (X = xi ), where E(X) = xi pX
i
i=1 i=1
n
X
= [xi − E(X)][xi − E(X)] · pX
i
i=1
Xn
= [xi − E(X)]2 · pX
i
i=1
= E[(X − E(X))2 ] (by def. of the expected value)
= V ar(X).

4
4. Cov (X, Y ) = Cov (Y, X) .

Proof: According to the definition of covariance, we can expand Cov(X, Y ) as follows:

Cov(X, Y ) = E[(X − E(X))(Y − E(Y ))]


n Xm n m
[xi − E(X)][yj − E(Y )] · pX,Y
X X X
= ij , where E(X) = xi p X
i and E(Y ) = yj pYj
i=1 j=1 i=1 j=1
m X n
[yj − E(Y )][xi − E(X)] · pX,Y
X
= ij
j=1 i=1

= E[(Y − E(Y ))(X − E(X))] (by def. of the expected value)


= Cov(Y, X). (by def. of the covariance)

5. Cov (a1 + b1 X, a2 + b2 Y ) = b1 b2 Cov (X, Y ) , where a1 , a2 , b1 , and b2 are some constants.

Proof: Using E(a1 + b1 X) = a1 + b1 E(X) and E(a2 + b2 Y ) = a2 + b2 E(Y ), we can expand


Cov (a1 + b1 X, a2 + b2 Y ) as follows:

Cov(X, Y ) = E[(a1 + b1 X − E(a1 + b1 X))(a2 + b2 Y − E(a2 + b2 Y ))]


= E[(a1 + b1 X − (a1 + b1 E(X)))(a2 + b2 Y − (a2 + b2 E(Y ))]
= E[(a1 − a1 + b1 X − b1 E(X))(a2 − a2 + b2 Y − b2 E(Y )]
= E[(b1 X − b1 E(X))(b2 Y − b2 E(Y )]
= E[b1 (X − E(X)) · b2 (Y − E(Y ))]
= E[b1 b2 (X − E(X))(Y − E(Y ))]
n X m
b1 b2 (xi − E(X))(yj − E(Y )) · pX,Y
X
= ij
i=1 j=1
n X
m
[xi − E(X)][yj − E(Y )] · pX,Y
X
= b1 b2 ij (by using (1))
i=1 j=1

= b1 b2 Cov(X, Y ).

6. If X and Y are independent, then Cov (X, Y ) = 0.

Proof: If X and Y are independent, by definition of stochastic independence, P (X = xi , Y =


yj ) = P (X = xi )P (Y = yj ) = pX Y
i pj for any i = 1, ..., n and j = 1, ..., m. Then, we may

5
expand Cov (X, Y ) as follows.

Cov(X, Y ) = E[(X − E(X))(Y − E(Y ))]


Xn Xm
= [xi − E(X)][yj − E(Y )] · P (X = xi , Y = yj )
i=1 j=1
Xn X m
= [xi − E(X)][yj − E(Y )]pX Y
i pj
i=1 j=1

because X and Y are independent


n X
X m
= {[xi − E(X)]pX Y
i }{[yj − E(Y )]pj }
i=1 j=1
 
n
X Xm 
= [xi − E(X)]pX
i [yj − E(Y )]pYj (5)
 
i=1 j=1
Pm
because we can move [xi − E(X)]pX
i outside of j=1
because [xi − E(X)]pX
i does not depend on the index j’s
 (
m n
)
X  X
= [yj − E(Y )]pYj [xi − E(X)]pX i (6)
 
j=1 i=1
nP o
m Y outside of ni=1
P
because we can move j=1 j[y − E(Y )]pj
nP o
m Y
because j=1 [y j − E(Y )]pj does not depend on the index i’s
( n  
n m m
)
X X X X 
= xi p X
i − E(X)p X
i · y p
j j
Y
− E(Y )p Y
j
 
i=1 i=1 j=1 j=1
( n
)  m

X  X 
= E(X) − E(X)pX i · E(Y ) − E(Y )p Y
j
 
i=1 j=1

by definition of E(X) and E(Y )


( n
)  m

X  X 
= E(X) − E(X) pX
i · E(Y ) − E(Y ) p Y
j
 
i=1 j=1
Pn Pm
because we can move E(X) and E(Y ) outside of i=1 and j=1 , respectively
= {E(X) − E(X) · 1} · {E(Y ) − E(Y ) · 1}
= 0 · 0 = 0.

Equation (6): This is similar to equation (4). Please consider the case of n = m = 2 and
convince yourself that (6) holds.

7. V ar (X + Y ) = V ar (X) + V ar (Y ) + 2Cov (X, Y ).

Proof: By the definition of variance,

V ar(X + Y ) = E[(X + Y − E(X + Y ))2 ].

6
Then,

V ar(X + Y ) = E[(X + Y − E(X + Y ))2 ]


= E[((X − E(X)) + (Y − E(Y )))2 ]
= E[(X − E(X))2 + (Y − E(Y ))2 + 2(X − E(X))(Y − E(Y ))]
because for any a and b, (a + b)2 = a2 + b2 + 2ab
= E[(X − E(X))2 ] + E[(Y − E(Y ))2 ] + 2E[(X − E(X))(Y − E(Y ))] (by using (2))
= V ar(X) + V ar(Y ) + 2Cov(X, Y )
by definition of variance and covariance

8. V ar (X − Y ) = V ar (X) + V ar (Y ) − 2Cov (X, Y ).

Proof: The proof of V ar (X − Y ) = V ar (X) + V ar (Y ) − 2Cov (X, Y ) is similar to the proof


of V ar (X + Y ) = V ar (X) + V ar (Y ) + 2Cov (X, Y ). First, we may show that E(X − Y ) =
E(X) − E(Y ). Then,

V ar(X − Y ) = E[(X − Y − E(X − Y ))2 ]


= E[((X − E(X)) − (Y − E(Y )))2 ]
= E[(X − E(X))2 + (Y − E(Y ))2 − 2(X − E(X))(Y − E(Y ))]
= E[(X − E(X))2 ] + E[(Y − E(Y ))2 ] − 2E[(X − E(X))(Y − E(Y ))] (by using (2))
= V ar(X) + V ar(Y ) − 2Cov(X, Y )
p p
9. Define W = (X −E(X))/ V ar(X) and Z = (Y −E(Y ))/ V ar(Y ). Show that Cov(W, Z) =
Corr(X, Z).

Proof: Expanding Cov(W, Z), we have

Cov(W, Z) = E[(W − E(W ))(Z − E(Z))]


= E[W Z] (because E[W ] = E[Z] = 0)
( )
X − E(X) Y − E(Y )
=E p ·p
V ar(X) V ar(Y )
by definition of W and Z
( )
1 1
=E p ·p · [X − E(X)]E[Y − E(Y )]
V ar(X) V ar(Y )
1 1
=p ·p · E {[X − E(X)]E[Y − E(Y )]} (by using (1))
V ar(X) V ar(Y )
because both √ 1 and √ 1 are constant
V ar(X) V ar(Y )
E {[X − E(X)]E[Y − E(Y )]}
= p p
V ar(X) V ar(Y )
Cov(X, Y )
=p p (by definition of covariance)
V ar(X) V ar(Y )
= Corr(X, Y ) (by definition of correlation coefficient)

7
10. Let b be a constant. Show that E[(X − b)2 ] = E(X 2 ) − 2bE(X) + b2 . What is the value of b
that gives the minimum value of E[(X − b)2 ]?

Answer: Because (X − b)2 = X 2 − 2bX + b2 , we have

E[(X − b)2 ] = E[X 2 − 2bX + b2 ] = E[X 2 ] − 2bE(X) + b2 .

Noting that E[X 2 ]−2bE(X)+b2 is a quadratic convex function of b, we may find the minimum

by differentiating E[(X − b)2 ] with respect to b and set ∂b E[(X − b)2 ] = 0, i.e.,


E[(X − b)2 ] = −2E(X) + 2b = 0,
∂b
and, therefore, setting the value of b equal to

b = E(X)

minimizes E[(X − b)2 ].

11. Let {xi : i = 1, . . . , n} and {yi : i = 1, . . . , n} be two sequences. Define the averages
n
1X
x̄ = xi ,
n
i=1
n
1X
ȳ = yi .
n
i=1
Pn
(a) i=1 (xi − x̄) = 0.

Proof:
n
X n
X n
X
(xi − x̄) = xi − x̄
i=1 i=1 i=1
Xn
= xi − nx̄
i=1
because ni=1 x̄ = x̄ + x̄ + ... + x̄ = nx̄
P
Pn
xi
= n i=1 − nx̄
n Pn
xi
because ni=1 xi = nn ni=1 xi = n i=1
P P
n
= nx̄ − nx̄
Pn
xi
because x̄ = i=1
n
= 0.
Pn
− x̄)2 =
Pn
(b) i=1 (xi i=1 xi (xi − x̄).

8
Proof: We use the result of 2.(a) above.
n
X n
X
2
(xi − x̄) = (xi − x̄) (xi − x̄)
i=1 i=1
Xn n
X
= xi (xi − x̄) − x̄ (xi − x̄)
i=1 i=1
Xn Xn
= xi (xi − x̄) − x̄ (xi − x̄)
i=1 i=1
n
X
because x̄ is constant and does not depend on i’s = xi (xi − x̄) − x̄ · 0
i=1
because ni=1 (xi − x̄) = 0. as shown above
P
n
X
= xi (xi − x̄) .
i=1
Pn Pn Pn
(c) i=1 (xi − x̄) (yi − ȳ) = i=1 yi (xi − x̄) = i=1 xi (yi − ȳ).

Proof: The proof is similar to the proof of 2.(b) above.


n
X n
X n
X
(xi − x̄) (yi − ȳ) = (xi − x̄) yi − (xi − x̄) ȳ
i=1 i=1 i=1
Xn Xn
= (xi − x̄) yi − ȳ (xi − x̄)
i=1 i=1
Xn
= (xi − x̄) yi − ȳ · 0
i=1
Xn
= yi (xi − x̄) .
i=1

Also,
n
X n
X n
X
(xi − x̄) (yi − ȳ) = xi (yi − ȳ) − x̄ (yi − ȳ)
i=1 i=1 i=1
Xn Xn
= xi (yi − ȳ) − x̄ (yi − ȳ)
i=1 i=1
Xn
= xi (yi − ȳ) − x̄ · 0
i=1
Xn
= xi (yi − ȳ) .
i=1

Conditional Mean and Conditional Variance


Let X and Y be two discrete random variables. The set of possible values for X is {x1 , . . . , xn };
and the set of possible values for Y is {y1 , . . . , ym }. We may define the conditional probability

9
function of Y given X as

Y |X P (X = xi , Y = yj ) pX,Y
ij
pij = P (Y = yj |X = xi ) = = X ,
P (X = xi ) pi

where pX,Y
ij = P (X = xi , Y = yj ) and pX
i = P (X = xi ).
The conditional mean of Y given X = xi is given by
m m
Y |X
X X
EY [Y |X = xi ] = yj P (Y = yj |X = xi ) = yj pij ,
j=1 j=1

where the symbol EY indicates that the expectation is taken treating Y as a random variable. The
conditional variance of Y given X = xi is given by
m
Y |X
X
V ar(Y |X = xi ) = E[(Y − E[Y |X = xi ])2 ] = (yj − E[Y |X = xi ])2 pij .
j=1

The conditional mean of Y given X can be written as EY [Y |X] without specifying a value of X.
Then, EY [Y |X] is a random variable because the value of EY [Y |X] depends on a realization of X.
The following shows that the unconditional mean of Y is equal to the expected value of EY [Y |X]
where the expectation is taken with respect to X.

1. Show that EY [Y ] = EX [EY [Y |X]].

Pm Y |X
Proof: Because EY [Y |X = xi ] = j=1 yj pij , we have
n
X
EX [EY [Y |X]] = EY [Y |X = xi ]pX
i
i=1
n X m
Y |X
X
= ( yj pij )pX
i
i=1 j=1
n X
X m
pX,Y
ij
= yj pX
i
i=1 j=1
pX
i
n X m
yj pX,Y
X
= ij
i=1 j=1
m n
pX,Y
X X
= yj ij
j=1 i=1
Xm
= yj pYj = EY [Y ].
j=1

2. Let g(Y ) be some known function of Y . Show that EY [g(Y )] = EX [EY [g(Y )|X]].

10
Proof:
n
X
EX [EY [g(Y )|X]] = EY [g(Y )|X = xi ]pX
i
i=1
n X m
Y |X
X
= ( g(yj )pij )pX
i
i=1 j=1
n X
X m
pX,Y
ij
= g(yj ) pX
i
i=1 j=1
pX
i
n X m
g(yj )pX,Y
X
= ij
i=1 j=1
m n
pX,Y
X X
= g(yj ) ij
j=1 i=1
Xm
= g(yj )pYj = EY [g(Y )].
j=1

3. Let g(Y ) and h(X) be some known functions of Y and X, respectively. Show that E[g(Y )h(X)] =
EX [h(X)EY [g(Y )|X]].

Proof:
n
X
EX [h(X)EY [g(Y )|X]] = h(xi )EY [g(Y )|X = xi ]pX
i
i=1
n m
Y |X
X X
= h(xi )( g(yj )pij )pX
i
i=1 j=1
n
X m
X pX,Y
ij
= h(xi ) g(yj ) pX
i
i=1 j=1
pX
i
n X
m
g(yj )h(xi )pX,Y
X
= ij
i=1 j=1

= E[g(Y )h(X)]

4. Show that, if E[Y |X] = EY [Y ], then Cov(X, Y ) = 0.

Proof:

Cov(X, Y ) = E[(X − EX (X))(Y − EY (Y ))] (by definition of Covariance)


= EX {[EY |X [(X − EX (X))(Y − EY (Y ))|X]} (by Law of Iterated Expectation)
= EX {(X − EX (X))EY |X [Y − EY (Y )|X]} (X is “known” once conditioned on X)
= EX {(X − EX (X))[EY |X (Y |X) − EY (Y )]} (EY (Y ) is a constant)
= EX {(X − EX (X))[EY (Y ) − EY (Y )]} (E[Y |X] = EY [Y ])
= EX [(X − EX (X)) × 0] = 0

Alternative Proof (Please compare this proof with the above proof ): Let EX (X) =

11
1 X 1 Y Y |X
n xi p i and EY (Y ) = m yj pj . Define pji = Pr(Y = yj |X = xi ).

Cov(X, Y ) = E(X,Y ) [(X − EX (X))(Y − EY (Y ))] (by definition of Covariance)


n m
1 1 XX
= (xi − EX (X))(yj − EY (Y ))pX,Y
ij
nm
i=1 j=1
 

 

 
X,Y 
 
n m
p

1 X  1 X ij

= (xi − EX (X))(yj − EY (Y )) X pX
i (by Law of Iterated Expectation)
n  m p i

i=1 
 j=1
 | {z } 


 
Y |X 
≡pji
  

 
 



 
 


n   m m 
1 X   1 X Y |X 1 X Y |X

= (xi − EX (X)) yj pji −EY (Y ) pji pX
i
n   m m 
i=1 




 j=1 j=1 



 |
 {z } | {z }


=E[Y |X] =1
n
1 X
= {(xi − EX (X))[EY (Y ) − EY (Y )]} pX
i (E[Y |X] = EY [Y ])
n
i=1
n
1 X
= {(xi − EX (X)) × 0} pX
i =0
n
i=1

12

You might also like