Lecture 4
Lecture 4
Rao-Blackwell Theorem
Let 𝑋1 , … … … … . , 𝑋𝑛 be a random sample from the density 𝑓(. ; 𝜃), and let
𝑆1 = 𝑠1 (𝑋1 , … … … … . , 𝑋𝑛 ), … … … . . , 𝑆𝑘 = 𝑠𝑘 (𝑋1 , … … … … . , 𝑋𝑛 ) be a set of jointly
sufficient statistics. Let the statistic 𝑇 = 𝑡(𝑋1 , … … … … . , 𝑋𝑛 ) be an unbiased
estimator of 𝜏(𝜃). Define 𝑇 ′ by 𝑇 ′ = 𝐸[𝑇/𝑆1 , 𝑆2 , … … . . , 𝑆𝑘 ].
Then,
(i) 𝑇 ′ is a statistic, and it is a function of the sufficient statistics 𝑆1 , … . . , 𝑆𝑘 .
Write 𝑇 ′ = 𝑡 ′ (𝑆1 , 𝑆2 , … … . . , 𝑆𝑘 ).
(ii) 𝐸[𝑇 ′ ] = 𝜏(𝜃); that is, 𝑇 ′ is an unbiased estimator of 𝜏(𝜃).
(iii) 𝑣𝑎𝑟[𝑇 ′ ] ≤ 𝑣𝑎𝑟[𝑇] for every 𝜃, and var[𝑇 ′ ] < 𝑣𝑎𝑟[𝑇] for some 𝜃 unless
T is equal to 𝑇 ′ with probability 1.
Proof: (i) 𝑆1 , 𝑆2 , … … . . , 𝑆𝑘 are sufficient statistics; so the conditional distribution of
any statistic, in particular the statistic T, given 𝑆1 , 𝑆2 , … … . . , 𝑆𝑘 is independent of 𝜃;
hence 𝑇 ′ = 𝐸[𝑇/𝑆1 , 𝑆2 , … … . . , 𝑆𝑘 ] is independent of 𝜃, and so 𝑇 ′ is a statistic which
is obviously a function of 𝑆1 , 𝑆2 , … … . . , 𝑆𝑘 .
(ii) 𝐸[𝑇 ′ ] = 𝐸[𝐸[𝑇/𝑆1 , 𝑆2 , … … . . , 𝑆𝑘 ]] = 𝐸[𝑇] = 𝜏(𝜃) [since E(Y)=E[E(Y/X)]
that is, 𝑇 ′ is an unbiased estimator of 𝜏(𝜃).
(iii) We can write
var[T]= 𝐸[(𝑇 − 𝐸[𝑇 ′ ])2 ] = 𝐸[(𝑇 − 𝑇 ′ + 𝑇 ′ − 𝐸[𝑇 ′ ])2 ]
= 𝐸[(𝑇 − 𝑇 ′ )2 ] + 2𝐸[(𝑇 − 𝑇 ′ )(𝑇 ′ − 𝐸[𝑇 ′ ])] + 𝑣𝑎𝑟[𝑇 ′ ]
But
𝐸[(𝑇 − 𝑇 ′ )(𝑇 ′ − 𝐸[𝑇 ′ ])] = 𝐸[𝐸[(𝑇 − 𝑇 ′ )(𝑇 ′ − 𝐸[𝑇 ′ ])/𝑆1 , 𝑆2 , … … . . , 𝑆𝑘 ]]
and
𝐸[(𝑇 − 𝑇 ′ )(𝑇 ′ − 𝐸[𝑇 ′ ])/𝑆1 = 𝑠1 , … … . . , 𝑆𝑘 = 𝑠𝑘 ]
= {𝑡 ′ (𝑠1 , … … … , 𝑠𝑘 ) − 𝐸[𝑇 ′ ]}𝐸[(𝑇 − 𝑇 ′ )/𝑆1 = 𝑠1 , … … . . , 𝑆𝑘 = 𝑠𝑘 ]
= {𝑡 ′ (𝑠1 , … … … , 𝑠𝑘 ) − 𝐸[𝑇 ′ ]}(𝐸[𝑇/𝑆1 = 𝑠1 , … … . . , 𝑆𝑘 = 𝑠𝑘 ]
−𝐸[𝑇 ′ /𝑆1 = 𝑠1 , … … . . , 𝑆𝑘 = 𝑠𝑘 ])
= {𝑡 ′ (𝑠1 , … … … , 𝑠𝑘 ) − 𝐸[𝑇 ′ ]}[𝑡 ′ (𝑠1 , … … … , 𝑠𝑘 ) − 𝑡 ′ (𝑠1 , … … … , 𝑠𝑘 )]
=0
and therefore
𝑣𝑎𝑟[𝑇] = 𝐸[(𝑇 − 𝑇 ′ )2 ] + 𝑣𝑎𝑟[𝑇 ′ ] ≥ 𝑣𝑎𝑟[𝑇 ′ ]
Note that 𝑣𝑎𝑟[𝑇] > 𝑣𝑎𝑟[𝑇 ′ ] unless T equals 𝑇 ′ with probability 1.
For many applications (particularly where the density involved has only one
unknown parameter) there will exist a single sufficient statistic, say 𝑆 =
𝑠(𝑋1 , … … … … . , 𝑋𝑛 ), which would then be used in place of the jointly sufficient set
of statistics 𝑆1 , 𝑆2 , … … . . , 𝑆𝑘 . What the theorem says is that, given an unbiased
estimator, another unbiased estimator that is a function of sufficient statistics can be
derived and it will not have larger variance.
Problem
Let 𝑋1 , 𝑋2 , … … … . , 𝑋𝑛 be a random sample from the Bernoulli density 𝑓(𝑥; 𝜃) =
𝜃 𝑥 (1 − 𝜃)1−𝑥 for x= 0 or 1. Let 𝑋1 is an unbiased estimator of 𝜏(𝜃) = 𝜃. Let ∑ 𝑋1
is a sufficient statistic. Show that 𝐸[𝑋1 / ∑ 𝑋1 ] is a statistic and an unbiased estimator
of 𝜃 with no larger variance than 𝑇 = 𝑋1 .
Solution:
We first find the conditional distribution of 𝑋1 given ∑ 𝑋1 = 𝑠. 𝑋1 takes on at most
the two values 0 and 1.
Now
𝑛−𝑠 𝑠 𝑠
𝐸 [𝑋1 / ∑ 𝑋𝑖 = 𝑠] = 0. + 1. =
𝑛 𝑛 𝑛
hence,
′ ∑𝑛
𝑖=1 𝑋𝑖
𝑇 = is a statistic
𝑛
𝜃(1−𝜃)
The variance of 𝑋1 is 𝜃(1 − 𝜃), and the variance of 𝑇 ′ is ; so for n>1 the
𝑛
variance of 𝑇 ′ is actually smaller than the variance of 𝑇 = 𝑋1 .
𝜃 0 𝑒 −𝜃
(b) 𝐸[𝑊] = 𝑃(𝑊 = 1) = 𝑃(𝑋1 = 0) = = 𝑒 −𝜃
0!
(c) Since 𝑋1 + 𝑋2~ Poisson (2𝜃), we have
𝐸[𝑊/𝑋1 + 𝑋2 = 𝑦] = ∑ 𝑊 𝑃(𝑊 = 1/𝑋1 + 𝑋2 = 𝑦) = 𝑃(𝑊 = 1/𝑋1 + 𝑋2 = 𝑦)
= 𝑃(𝑋1 = 0/𝑋1 + 𝑋2 = 𝑦)
𝑃(𝑋1 =0 𝑎𝑛𝑑 𝑋1 +𝑋2 =𝑦)
=
𝑃(𝑋1 +𝑋2 =𝑦)
𝑃(𝑋1 =0 𝑎𝑛𝑑 𝑋2 =𝑦)
=
𝑃(𝑋1 +𝑋2 =𝑦)
𝑃(𝑋1 =0) 𝑃(𝑋2 =𝑦)
=
𝑃(𝑋1 +𝑋2 =𝑦)
𝑒−𝜃 𝜃𝑦
𝑒 −𝜃 ( )
𝑦!
=
𝑒 −2𝜃 (2𝜃)𝑦 /𝑦!
𝜃𝑦 1 𝑦
= =( )
(2𝜃)𝑦 2
(d) Since W is an unbiased estimator of 𝑒 −𝜃 and 𝑋1 + 𝑋2 is sufficient for 𝜃 (and
consequently 𝑒 −𝜃 ), the Rao-Blackwell Theorem implies that
1 𝑋1 +𝑋2
𝑇 ′ (𝑋1 + 𝑋2 ) = 𝐸[𝑊/𝑋1 + 𝑋2 ] = ( )
2
is better uniformly unbiased estimator of 𝑒 −𝜃 .