Chapter #6 – Functions of Random Variables Question #2:

advertisement
Chapter #6 – Functions of Random Variables
Question #2: Let 𝑋~π‘ˆπ‘πΌπΉ(0,1). Use the CDF technique to find the PDF of the following
random variables: a) π‘Œ = 𝑋1/4 , b) π‘Š = 𝑒 −𝑋 , c) 𝑍 = 1 − 𝑒 −𝑋 , and d) π‘ˆ = 𝑋(1 − 𝑋).
1 𝑖𝑓 π‘₯ ∈ (0,1)
a) Since 𝑋~π‘ˆπ‘πΌπΉ(0,1), we know that the density function is 𝑓𝑋 (π‘₯) = {
0 π‘œπ‘‘β„Žπ‘’π‘Ÿπ‘€π‘–π‘ π‘’
0
while the distribution function is 𝐹𝑋 (π‘₯) = {π‘₯
1
𝑖𝑓 π‘₯ ∈ (−∞, 0]
𝑖𝑓 π‘₯ ∈ (0,1) . We can then use the
𝑖𝑓 π‘₯ ∈ [1, ∞)
CDF technique to find πΉπ‘Œ (𝑦) = 𝑃(π‘Œ ≤ 𝑦) = 𝑃(𝑋1/4 ≤ 𝑦) = 𝑃(𝑋 ≤ 𝑦 4 ) = 𝐹𝑋 (𝑦 4 ), so
𝑑
𝑑
that π‘“π‘Œ (𝑦) = 𝑑𝑦 πΉπ‘Œ (𝑦) = 𝑑𝑦 𝐹𝑋 (𝑦 4 ) = 𝑓𝑋 (𝑦 4 )4𝑦 3 = (1)4𝑦 3 . Since 0 < π‘₯ < 1, the
bounds are 0 < 𝑦 4 < 1 or 0 < 𝑦 < 1. Therefore, π‘“π‘Œ (𝑦) = {4𝑦
0
3
𝑖𝑓 𝑦 ∈ (0,1).
π‘œπ‘‘β„Žπ‘’π‘Ÿπ‘€π‘–π‘ π‘’
b) We have πΉπ‘Š (𝑀) = 𝑃(π‘Š ≤ 𝑀) = 𝑃(𝑒 −𝑋 ≤ 𝑀) = 𝑃(𝑋 ≥ −𝑙𝑛(𝑀)) = 1 − 𝐹𝑋 (−𝑙𝑛(𝑀)),
𝑑
𝑑
1
1
so that π‘“π‘Š (𝑀) = 𝑑𝑀 πΉπ‘Š (𝑀) = 𝑑𝑀 [1 − 𝐹𝑋 (− 𝑙𝑛(𝑀))] = −𝑓𝑋 (− 𝑙𝑛(𝑀)) (− 𝑀) = 𝑀. Since
0 < π‘₯ < 1, the bounds are 0 < − 𝑙𝑛(𝑀) < 1 or 𝑒 −1 < 𝑀 < 1. The probability density
1
𝑖𝑓 𝑀 ∈ (𝑒 −1 , 1)
.
π‘œπ‘‘β„Žπ‘’π‘Ÿπ‘€π‘–π‘ π‘’
function of the random variable π‘Š is therefore π‘“π‘Š (𝑀) = {𝑀
0
c) By the CDF technique, we have that the 𝐹𝑍 (𝑧) = 𝑃(𝑍 ≤ 𝑧) = 𝑃(1 − 𝑒 −𝑋 ≤ 𝑧) =
𝑃(𝑒 −𝑋 ≥ 1 − 𝑧) = 𝑃(𝑋 ≤ −𝑙 𝑛(1 − 𝑧)) = 𝐹𝑋 (−𝑙 𝑛(1 − 𝑧)), so we have that 𝑓𝑍 (𝑧) =
𝑑
𝐹 (𝑧)
𝑑𝑧 𝑍
𝑑
−1
1
= 𝑑𝑧 𝐹𝑋 (−𝑙 𝑛(1 − 𝑧)) = 𝑓𝑋 (− 𝑙𝑛(1 − 𝑧)) (− 1−𝑧) = 1−𝑧. Since 0 < π‘₯ < 1, the
bounds are 0 < −𝑙 𝑛(1 − 𝑧) < 1 or 0 < 𝑧 < 1 − 𝑒 −1 . Therefore, the probability
1
density function of this random variable is 𝑓𝑍 (𝑧) = { 1−𝑧
0
𝑖𝑓 𝑧 ∈ (0,1 − 𝑒 −1 )
π‘œπ‘‘β„Žπ‘’π‘Ÿπ‘€π‘–π‘ π‘’
.
d) We will use Theorem 6.3.2 for this question. Suppose that 𝑋 is a continuous random
variable with density 𝑓𝑋 (π‘₯) and assume that π‘Œ = 𝑒(𝑋) is a one-to-one transformation
𝑑
with inverse π‘₯ = 𝑀(𝑦). If 𝑑𝑦 𝑀(𝑦) is continuous and nonzero, the density of π‘Œ is given
𝑑
by π‘“π‘Œ (𝑦) = 𝑓𝑋 (𝑀(𝑦)) |𝑑𝑦 𝑀(𝑦)|. Here, the transformation is 𝑒 = π‘₯(1 − π‘₯) = −π‘₯ 2 + π‘₯
and, based on the y-values of the graph over the x-values of (0,1), the range over
1
which the density of π‘ˆ is defined is (0, 4). Since this transformation is not one-to-one,
1
1
we must partition the interval (0,1) into two parts: (0, 2) and (2 , 1). Then solve the
transformation by completing the square to obtain 𝑒 = −π‘₯ 2 + π‘₯ → −𝑒 = π‘₯ 2 − π‘₯ →
1
1
− 𝑒 = π‘₯2 − π‘₯ + 4 →
4
1
1 2
1
1
1
Since π‘₯ = 𝑀(𝑒) = 2 ± √4 − 𝑒, we have 𝑀
1
1
1
1
− 𝑒 = (π‘₯ − 2) → π‘₯ − 2 = ±√4 − 𝑒 → π‘₯ = 2 ± √4 − 𝑒.
4
1
1
′ (𝑒)
1 1
= ± 2 (4 − 𝑒)
−
1
2
(−1) =
±1
1
4
. Thus,
2√ −𝑒
1
1
π‘“π‘ˆ (𝑒) = 𝑓𝑋 (2 + √4 − 𝑒) |𝑀 ′ (𝑒)| + 𝑓𝑋 (2 − √4 − 𝑒) |𝑀 ′ (𝑒)| = (1 + 1)|π‘₯ ′ (𝑒)| =
−
1
so we can conclude that the density function is π‘“π‘ˆ (𝑒) = {(4 − 𝑒)
0
1
2
,
1
√ −𝑒
4
1
𝑖𝑓 𝑒 ∈ (0, 4).
π‘œπ‘‘β„Žπ‘’π‘Ÿπ‘€π‘–π‘ π‘’
Question #3: The measured radius 𝑅 of a circle has PDF 𝑓𝑅 (π‘Ÿ) = 6π‘Ÿ(1 − π‘Ÿ) if π‘Ÿ ∈ (0,1) and
𝑓𝑅 (π‘Ÿ) = 0 otherwise. Find the distribution of a) the circumference and b) area of the circle.
a) The circumference of a circle is given by 𝐢 = 2πœ‹π‘…, so by the CDF technique we have
𝑐
𝑐
that 𝐹𝐢 (𝑐) = 𝑃(𝐢 ≤ 𝑐) = 𝑃(2πœ‹π‘… ≤ 𝑐) = 𝑃 (𝑅 ≤ 2πœ‹) = 𝐹𝑅 (2πœ‹). The density is thus
𝑑
𝑑
𝑐
𝑐
1
𝑐
𝑐
1
𝑓𝐢 (𝑐) = 𝑑𝑐 𝐹𝐢 (𝑐) = 𝑑𝑐 𝐹𝑅 (2πœ‹) = 𝑓𝑅 (2πœ‹) (2πœ‹) = 6 (2πœ‹) (1 − 2πœ‹) (2πœ‹) = β‹― =
6𝑐(2πœ‹−𝑐)
(2πœ‹)3
.
𝑐
Since 0 < π‘Ÿ < 1 we have 0 < 2πœ‹ < 1 so that 0 < 𝑐 < 2πœ‹. Therefore, the probability
density function of the circumference is given by 𝑓𝐢 (𝑐) = {
6𝑐(2πœ‹−𝑐)
(2πœ‹)3
𝑖𝑓 𝑐 ∈ (0,2πœ‹)
0
π‘œπ‘‘β„Žπ‘’π‘Ÿπ‘€π‘–π‘ π‘’
.
b) The area is given by 𝐴 = πœ‹π‘… 2 , so we have 𝐹𝐴 (π‘Ž) = 𝑃(𝐴 ≤ π‘Ž) = 𝑃(πœ‹π‘… 2 ≤ π‘Ž) =
π‘Ž
π‘Ž
π‘Ž
π‘Ž
π‘Ž
π‘Ž
𝑃 (𝑅 2 ≤ πœ‹) = 𝑃 (|𝑅| ≤ ±√πœ‹) = 𝑃 (−√πœ‹ ≤ 𝑅 ≤ √πœ‹) = 𝐹𝑅 (√πœ‹) − 𝐹𝑅 (−√πœ‹).
Thus,
𝑑
𝑑
π‘Ž
π‘Ž
3(√πœ‹−√π‘Ž)
.
πœ‹ 3/2
we have 𝑓𝐴 (π‘Ž) = π‘‘π‘Ž 𝐹𝐴 (π‘Ž) = π‘‘π‘Ž [𝐹𝑅 (√πœ‹) − 𝐹𝑅 (−√πœ‹)] = β‹― =
π‘Ž
0 < π‘Ÿ < 1, then 0 < √πœ‹ < 1 so that 0 < π‘Ž < πœ‹ and 𝑓𝐴 (π‘Ž) = {
3(√πœ‹−√π‘Ž)
πœ‹ 3/2
0
Since we have
𝑖𝑓 π‘Ž ∈ (0, πœ‹).
π‘œπ‘‘β„Žπ‘’π‘Ÿπ‘€π‘–π‘ π‘’
1
Question #10: Suppose 𝑋 has density 𝑓𝑋 (π‘₯) = 2 𝑒 −|π‘₯| for all π‘₯ ∈ ℝ. a) Find the density of the
random variable π‘Œ = |𝑋|. b) If π‘Š = 0 when 𝑋 ≤ 0 and π‘Š = 1 when 𝑋 > 0 find the CDF of π‘Š.
a) We have πΉπ‘Œ (𝑦) = 𝑃(π‘Œ ≤ 𝑦) = 𝑃(|𝑋| < 𝑦) = 𝑃(−𝑦 ≤ 𝑋 ≤ 𝑦) = 𝐹𝑋 (𝑦) − 𝐹𝑋 (−𝑦), so
that we obtain π‘“π‘Œ (𝑦) =
1
2
𝑑
𝐹 (𝑦)
𝑑𝑦 π‘Œ
=
𝑑
𝑑𝑦
[𝐹𝑋 (𝑦) − 𝐹𝑋 (−𝑦)] = 𝑓𝑋 (𝑦)(1) − 𝑓𝑋 (−𝑦)(−1) =
1
𝑒 −|𝑦| + 2 𝑒 −|−𝑦| = 𝑒 −|𝑦| . Since the transformation was an absolute value function and
we have the bounds −∞ < π‘₯ < ∞, the bounds become 0 < 𝑦 < ∞. This allows us to
write the probability density function of π‘Œ as π‘“π‘Œ (𝑦) = {
1
𝑒 −𝑦
0
𝑖𝑓 𝑦 ∈ (0, ∞)
.
π‘œπ‘‘β„Žπ‘’π‘Ÿπ‘€π‘–π‘ π‘’
1
b) We see that 𝑃(π‘Š = 0) = 2 and 𝑃(π‘Š = 1) = 2 since 𝑓𝑋 (π‘₯) is symmetric. This allows us
0
𝑖𝑓 𝑀 ∈ (−∞, 0)
1
𝑖𝑓 𝑀 ∈ [0,1)
1
𝑖𝑓 𝑀 ∈ [1, ∞)
to write the cumulative distribution function as πΉπ‘Š (𝑀) = { 2
.
1
Question #13: Suppose 𝑋 has density 𝑓𝑋 (π‘₯) = 24 π‘₯ 2 for π‘₯ ∈ (−2,4) and 𝑓𝑋 (π‘₯) = 0 otherwise.
Find the probability density function of the random variable π‘Œ = 𝑋 2 .
ο‚·
We will use Theorem 6.3.2 for this question: Suppose that 𝑋 is a continuous random
variable with density 𝑓𝑋 (π‘₯) and assume that π‘Œ = 𝑒(𝑋) is a one-to-one transformation
𝑑
with inverse π‘₯ = 𝑀(𝑦). If 𝑑𝑦 𝑀(𝑦) is continuous and nonzero, the density of π‘Œ is given
𝑑
by π‘“π‘Œ (𝑦) = 𝑓𝑋 (𝑀(𝑦)) |𝑑𝑦 𝑀(𝑦)|. Here, the transformation is 𝑒 = π‘₯ 2 and, based on the
y-values of the graph over the x-values of (0,1), the domain over which the density of
π‘Œ is defined is (0,16). Solving the transformation then gives that π‘₯ = 𝑀(𝑦) = ±√𝑦 so
1
that 𝑀 ′ (𝑦) = ± 2 𝑦. We must consider two cases in the interval (0,16): over (0,4) the
√
transformation is not one-to-one and over (4,16) it is one-to-one. The density is thus
𝑓𝑋 (−√𝑦)|𝑀 ′ (𝑦)| + 𝑓𝑋 (√𝑦)|𝑀 ′ (𝑦)| 𝑖𝑓 𝑦 ∈ (0,4)
π‘“π‘Œ (𝑦) = {𝑓 (√𝑦)|𝑀 ′ (𝑦)|
𝑋
0
𝑖𝑓 𝑦 ∈ [4,16) =
π‘œπ‘‘β„Žπ‘’π‘Ÿπ‘€π‘–π‘ π‘’
√𝑦
24
{√𝑦
𝑖𝑓 𝑦 ∈ (0,4)
𝑖𝑓 𝑦 ∈ [4,16)
48
0
.
π‘œπ‘‘β„Žπ‘’π‘Ÿπ‘€π‘–π‘ π‘’
Question #16: Let 𝑋1 and 𝑋2 be independent random variables each having density function
1
𝑓𝑋𝑖 (π‘₯) = π‘₯ 2 for π‘₯𝑖 ∈ [1, ∞) and 𝑓𝑋𝑖 (π‘₯) = 0 otherwise. a) Find the joint PDF of π‘ˆ = 𝑋1 𝑋2 and
𝑉 = 𝑋1. b) Find the marginal probability density function of the random variable π‘ˆ.
a) Since 𝑋1 and 𝑋2 are independent, their joint density is simply the product of their
1
1
1
2
marginal densities, so 𝑓𝑋1 𝑋2 (π‘₯1 , π‘₯2 ) = (π‘₯ 2 ) (π‘₯ 2 ) = (π‘₯
1
2
1 π‘₯2 )
whenever we have that
(π‘₯1 , π‘₯2 ) ∈ [1, ∞) × [1, ∞) and zero otherwise. We will use Theorem 6.3.6, which says
that if 𝑒 = 𝑓(π‘₯1 , π‘₯2 ) and 𝑣 = 𝑓(π‘₯1 , π‘₯2 ) and we can solve uniquely for π‘₯1 and π‘₯2 , then
πœ•π‘₯1
we have π‘“π‘ˆπ‘‰ (𝑒, 𝑣) = 𝑓𝑋1 𝑋2 (π‘₯1 (𝑒, 𝑣), π‘₯2 (𝑒, 𝑣))|𝐽|, where 𝐽 =
πœ•π‘’
𝑑𝑒𝑑 [πœ•π‘₯
2
πœ•π‘₯1
πœ•π‘£
].
πœ•π‘₯2
πœ•π‘’
𝑒
Here, we
πœ•π‘£
𝑒
have that 𝑣 = π‘₯1 so π‘₯1 = 𝑣 and 𝑒 = π‘₯1 π‘₯2 so π‘₯2 = π‘₯ = 𝑣 , so we can calculate the
1
0
1
1
Jacobian as 𝐽 = 𝑑𝑒𝑑 [ 1 − 𝑒 ] = − 𝑣. We can therefore find the joint density as
𝑣
𝑣2
𝑒
1
π‘“π‘ˆπ‘‰ (𝑒, 𝑣) = 𝑓𝑋1 𝑋2 (𝑣, 𝑣 ) |− 𝑣| = (
1
1
𝑒 2
(𝑣⋅ )
𝑣
1
) (𝑣) = 𝑒2 𝑣 if 1 < 𝑣 < 𝑒 < ∞. We can find this
region of integration by substituting into the constraints of 𝑓𝑋1 𝑋2 (π‘₯1 , π‘₯2 ). The first is
that 1 < π‘₯1 < ∞ so 1 < 𝑣 < ∞ while the second is 1 < π‘₯2 < ∞ so 1 <
𝑒
𝑣
< ∞, which
reduces to 𝑣 < 𝑒 < ∞. Combining these gives the required bounds 1 < 𝑣 < 𝑒 < ∞.
𝑒
𝑒 1
b) We have π‘“π‘ˆ (𝑒) = ∫1 π‘“π‘ˆπ‘‰ (𝑒, 𝑣)𝑑𝑣 = ∫1
1
𝑒
𝑙𝑛(𝑒)
1
𝑒2
𝑑𝑣 = [𝑒2 𝑙𝑛(𝑣)] =
𝑒2 𝑣
if 1 < 𝑒 < ∞.
Question #18: Let 𝑋 and π‘Œ have joint density function π‘“π‘‹π‘Œ (π‘₯, 𝑦) = 𝑒 −𝑦 for 0 < π‘₯ < 𝑦 < ∞
and π‘“π‘‹π‘Œ (π‘₯, 𝑦) = 0 otherwise. a) Then find the joint density function of 𝑆 = 𝑋 + π‘Œ and 𝑇 = 𝑋.
b) Find the marginal density function of 𝑆. c) Find the marginal density function of 𝑇.
a) We have that 𝑑 = π‘₯ so π‘₯ = 𝑑 and 𝑠 = π‘₯ + 𝑦 so 𝑦 = 𝑠 − π‘₯ = 𝑠 − 𝑑, so we can calculate
πœ•π‘₯
the Jacobian as 𝐽 =
πœ•π‘₯
πœ•π‘ 
𝑑𝑒𝑑 [πœ•π‘¦
πœ•π‘‘
]
πœ•π‘¦
πœ•π‘ 
πœ•π‘‘
0
= 𝑑𝑒𝑑 [
1
1
] = −1. We can therefore find the joint
−1
𝑠
density as 𝑓𝑆𝑇 (𝑠, 𝑑) = π‘“π‘‹π‘Œ (𝑑, 𝑠 − 𝑑)|−1| = (𝑒 −(𝑠−𝑑) )(1) = 𝑒 𝑑−𝑠 if 0 < 𝑑 < 2 < ∞. We
can find these bounds by substituting into the bounds 0 < π‘₯ < 𝑦 < ∞ with our solved
𝑠
transformations, which gives 0 < 𝑑 < 𝑠 − 𝑑 < ∞ or 0 < 2𝑑 < 𝑠 < ∞ or 0 < 𝑑 < 2 < ∞.
𝑠/2
b) We have that 𝑓𝑆 (𝑠) = ∫0
𝑠/2
𝑒 −𝑠 [𝑒 𝑑 ]0
𝑠/2 𝑑−𝑠
𝑓𝑆𝑇 (𝑠, 𝑑)𝑑𝑑 = ∫0
𝑒
𝑠/2 𝑑 −𝑠
𝑑𝑑 = ∫0
𝑠/2
𝑒 𝑒 𝑑𝑑 = [𝑒 𝑑 𝑒 −𝑠 ]0
=
= 𝑒 −𝑠 (𝑒 𝑠/2 − 1) = if 0 < 𝑠 < ∞.
∞
∞
∞
c) We have that 𝑓𝑇 (𝑑) = ∫2𝑑 𝑓𝑆𝑇 (𝑠, 𝑑)𝑑𝑠 = ∫2𝑑 𝑒 𝑑−𝑠 𝑑𝑠 = ∫2𝑑 𝑒 𝑑 𝑒 −𝑠 𝑑𝑠 = [−𝑒 𝑑 𝑒 −𝑠 ]∞
2𝑑 =
𝑑
−2𝑑
−𝑒 𝑑 [𝑒 −𝑠 ]∞
) = 𝑒 −𝑑 if 0 < 𝑑 < ∞. Note that we have omitted the steps
2𝑑 = −𝑒 (0 − 𝑒
where the infinite limit of integration is replaced by a parameter and a limit to infinity
with that parameter is evaluated to show that it goes to zero.
Question #21: Let 𝑋 and π‘Œ have joint density π‘“π‘‹π‘Œ (π‘₯, 𝑦) = 2(π‘₯ + 𝑦) for 0 < π‘₯ < 𝑦 < 1 and
π‘“π‘‹π‘Œ (π‘₯, 𝑦) = 0 otherwise. a) Find the joint probability density function of 𝑆 = 𝑋 and 𝑇 = π‘‹π‘Œ.
b) Find the marginal probability density function of the random variable 𝑇.
𝑑
𝑑
a) We have that 𝑠 = π‘₯ so π‘₯ = 𝑠 and 𝑑 = π‘₯𝑦 so 𝑦 = π‘₯ = 𝑠, so we can calculate the Jacobian
πœ•π‘₯
πœ•π‘ 
as 𝐽 = 𝑑𝑒𝑑 [πœ•π‘¦
πœ•π‘ 
πœ•π‘₯
πœ•π‘‘
]
πœ•π‘¦
1
= 𝑑𝑒𝑑 [− 𝑑
𝑠
πœ•π‘‘
𝑑
1
0
1
1 ] = . The joint probability density function is thus
𝑠
𝑠
𝑑
1
𝑑
𝑓𝑆𝑇 (𝑠, 𝑑) = π‘“π‘‹π‘Œ (𝑠, 𝑠) |𝑠 | = 2 (𝑠 + 𝑠) (𝑠 ) = 2 (1 + 𝑠2 ). The region is then found by
𝑑
substituting to 0 < π‘₯ < 𝑦 < 1, so we have 0 < 𝑠 < 𝑠 < 1 or 0 < 𝑠 2 < 𝑑 < 𝑠 < 1. This
can be visualized as the region between 𝑑 = 𝑠 and 𝑑 = 𝑠 2 on the 𝑠𝑑 plane.
√𝑑
b) The marginal probability density function is given by 𝑓𝑇 (𝑑) = ∫𝑑 𝑓𝑆𝑇 (𝑠, 𝑑)𝑑𝑠 =
√𝑑
2𝑑 √𝑑
𝑑
∫𝑑 2 (1 + 𝑠2 ) 𝑑𝑠 = [2𝑠 − 𝑠 ]
𝑑
= (2√𝑑 − 2√𝑑) − (2𝑑 − 2) = 2 − 2𝑑 if 0 < 𝑑 < 1.
Question #25: Let 𝑋1 , 𝑋2 , 𝑋3,𝑋4 be independent random variables. Assume that 𝑋2, 𝑋3, 𝑋4 are
each distributed Poisson with parameter 5 and the random variable π‘Œ = 𝑋1 + 𝑋2 + 𝑋3 + 𝑋4
is distributed Poisson with parameter 25. a) What is the distribution of the random variable
𝑋1? b) What is the distribution of the random variable π‘Š = 𝑋1 + 𝑋2?
a) We first note that while 𝑋1,𝑋2, 𝑋3, 𝑋4 are independent, they are not 𝑖𝑖𝑑 since only
𝑋2 , 𝑋3 , 𝑋4 ~𝑃𝑂𝐼(5) with 𝑋1 not being listed. Thus, we must use the unsimplified
formula 6.4.4, which says if 𝑋1 , … , 𝑋𝑛 are independent random variables with moment
generating functions 𝑀𝑋𝑖 (𝑑) and π‘Œ = 𝑋1 + β‹― + 𝑋𝑛 , then the moment generating
function of π‘Œ is π‘€π‘Œ (𝑑) = [𝑀𝑋1 (𝑑)] … [𝑀𝑋𝑛 (𝑑)]. We use the fact that if some 𝑋~𝑃𝑂𝐼(πœ‡),
then 𝑀𝑋 (𝑑) = 𝑒 πœ‡(𝑒
that
𝑑 −1)
to solve this problem. If we let π‘Œ = 𝑋1 + 𝑋2 + 𝑋3 + 𝑋4 , we have
π‘€π‘Œ (𝑑) = [𝑀𝑋1 (𝑑)][𝑀𝑋2 (𝑑)][𝑀𝑋3 (𝑑)][𝑀𝑋4 (𝑑)] = 𝑒 25(𝑒
Substituting gives [𝑀𝑋1 (𝑑)][𝑒 5(𝑒
[𝑀𝑋1 (𝑑)][𝑒 5(𝑒
𝑑 −1)
3
] = 𝑒 5(𝑒
𝑑 −1)
𝑑 −1)
][𝑒 5(𝑒
𝑑 −1)
][𝑒 5(𝑒
so that [𝑀𝑋1 (𝑑)] =
𝑑 −1)
𝑑 −1)
] = 𝑒 25(𝑒
𝑑
𝑒 25(𝑒 −1)
3
𝑑
[𝑒 5(𝑒 −1) ]
=
since
𝑑 −1)
π‘Œ~𝑃𝑂𝐼(25).
, which reduces to
𝑑
𝑒 25(𝑒 −1)
𝑑
𝑒 15(𝑒 −1)
= 𝑒 10(𝑒
𝑑 −1)
. This
is the moment generating function of a poisson 10 random variable, so 𝑋1 ~𝑃𝑂𝐼(10).
b) We have π‘€π‘Š (𝑑) = [𝑀𝑋1 (𝑑)][𝑀𝑋2 (𝑑)] = [𝑒 10(𝑒
𝑑 −1)
][𝑒 5(𝑒
𝑑 −1)
] = 𝑒 15(𝑒
𝑑 −1)
, which is the
moment generating function of a poisson 15 random variable, so π‘Š~𝑃𝑂𝐼(15). We can
see a general pattern here; if 𝑋𝑖 ~𝑃𝑂𝐼(πœ‡) for 𝑖 = 1, … , 𝑛 are independent random
variables and we define some 𝑍 = ∑π‘˜π‘–=1 𝑋𝑖 for π‘˜ ≤ 𝑛, then we have that 𝑍~𝑃𝑂𝐼(π‘˜πœ‡).
Chapter #6 – Functions of Random Variables
Question #17: Suppose that 𝑋1 and 𝑋2 denote a random sample of size 2 from a gamma
1
𝑋
distribution such that 𝑋𝑖 ~𝐺𝐴𝑀 (2, 2). Find the PDF of a) π‘Œ = √𝑋1 + 𝑋2 and b) π‘Š = 𝑋1.
2
π‘₯
1
a) We know that if some 𝑋~𝐺𝐴𝑀(πœƒ, πœ…), then 𝑓𝑋 (π‘₯) = πœƒπœ… Γ(πœ…) π‘₯ πœ…−1 𝑒 −πœƒ if π‘₯ > 0. Since 𝑋1
and 𝑋2 are independent, their joint density function is given by 𝑓𝑋1 𝑋2 (π‘₯1 , π‘₯2 ) =
−
1
𝑓𝑋1 (π‘₯1 )𝑓𝑋2 (π‘₯2 ) = (
1
√2Γ(2)
1
2
π‘₯1 𝑒
π‘₯
− 1
2
1
)(
1
√2Γ(2)
−
1
2
π‘₯2
1
π‘₯2 𝑒 − 2 ) = 2πœ‹
1
1
√ π‘₯1 √ π‘₯2
𝑒−
(π‘₯1 +π‘₯2 )
2
if π‘₯1 , π‘₯2 > 0,
1
since Γ (2) = √πœ‹. We have the transformation 𝑦 = √π‘₯1 + π‘₯2 and generate another
transformation 𝑀 = π‘₯1 , so we have π‘₯1 = 𝑀 and π‘₯2 = 𝑦 2 − 𝑀, which allows us to find
1
0
𝐽 = 𝑑𝑒𝑑 [
] = 2𝑦. Then the joint density of π‘Š and π‘Œ is given by π‘“π‘Šπ‘Œ (𝑀, 𝑦) =
−1 2𝑦
1 1
𝑓𝑋1 𝑋2 (𝑀, 𝑦 2 − 𝑀)|𝐽| = πœ‹
𝑦2
1
√𝑀 √𝑦 2 −𝑀
𝑒 − 2 𝑦 if 𝑀 > 0, 𝑦 2 − 𝑀 > 0; these bounds can be
∞
combined to give 0 < 𝑀 < 𝑦 2 . Finally, the density of π‘Œ is π‘“π‘Œ (𝑦) = ∫−∞ π‘“π‘Šπ‘Œ (𝑀, 𝑦) 𝑑𝑀 =
1
𝑦2
1
𝑦2
1
∫ 𝑦 √𝑀 √𝑦 2
πœ‹ 0
−𝑀
𝑦2
𝑒 − 2 𝑑𝑀 = β‹― = 𝑦𝑒 − 2 if 𝑦 > 0 and zero otherwise. The evaluation of
this integral has been omitted, but can be computed by two substitutions.
π‘₯
𝑧
b) We have 𝑀 = π‘₯1 and generate 𝑧 = π‘₯1 so that π‘₯1 = 𝑧 and π‘₯2 = 𝑀, which allows us to
2
1
calculate 𝐽 = det [ 1
𝑀
1 1
1
2πœ‹ √𝑧 √𝑧/𝑀
𝑒
𝑧+𝑧/𝑀
−
2
∞
𝑧
𝑀2
0
𝑧
𝑧
𝑧
(𝑧,
|𝐽|
− 𝑀2 ] = − 𝑀2 . Then we have π‘“π‘π‘Š 𝑀) = 𝑓𝑋1 𝑋2 (𝑧, 𝑀) =
1
1
= 2πœ‹ 𝑀3/2 𝑒 −
1
(𝑧+𝑧/𝑀)
2
∞
π‘“π‘Š (𝑀) = ∫−∞ π‘“π‘Šπ‘ (𝑀, 𝑧) 𝑑𝑧 = 2πœ‹ ∫0
1
if 𝑧, 𝑀 > 0, so that the density of π‘Š is given by
𝑒−
𝑀 3/2
(𝑧+𝑧/𝑀)
2
1
1
𝑑𝑧 = β‹― = πœ‹ (𝑀+1)
√𝑀
if 𝑀 > 0. The
evaluation of this integral has been omitted, but can be computed by substitution.
Question #26: Let 𝑋1 and 𝑋2 be independent negative binomial random variables such that
𝑋1 ~𝑁𝐡(π‘Ÿ1 , 𝑝) and 𝑋2 ~𝑁𝐡(π‘Ÿ2 , 𝑝). a) Find the MGF and distribution of π‘Œ = 𝑋1 + 𝑋2.
ο‚·
We use Theorem 6.4.3, which says that if the random variables 𝑋𝑖 are independent
with respective MGFs 𝑀𝑋𝑖 (𝑑), then the MGF of the random variable that is their sum
is simply the product of their respective MGFs. Also, if some discrete random variable
π‘Ÿ
𝑝𝑒 𝑑
𝑋~𝑁𝐡(π‘Ÿ, 𝑝), then 𝑀𝑋 (𝑑) = (1−π‘žπ‘’ 𝑑) . Therefore, the moment generating function of π‘Œ
𝑝𝑒 𝑑
is π‘€π‘Œ (𝑑) = [𝑀𝑋1 (𝑑)][𝑀𝑋2 (𝑑)] = (
1−π‘žπ‘’ 𝑑
π‘Ÿ1
) (
𝑝𝑒 𝑑
1−π‘žπ‘’ 𝑑
π‘Ÿ2
)
𝑝𝑒 𝑑
=(
1−π‘žπ‘’ 𝑑
π‘Ÿ1 +π‘Ÿ2
)
. This then allows to
determine the distribution of π‘Œ = 𝑋1 + 𝑋2, namely that π‘Œ~𝑁𝐡(π‘Ÿ1 + π‘Ÿ2 , 𝑝).
Question #27: Recall that π‘Œ~𝐿𝑂𝐺𝑁(πœ‡, 𝜎 2 ) if 𝑙𝑛(π‘Œ) ~𝑁(πœ‡, 𝜎 2 ). Assume that π‘Œπ‘– ~𝐿𝑂𝐺𝑁(πœ‡π‘– , πœŽπ‘–2 )
for 𝑖 = 1, … , 𝑛 are independent. Find the distribution functions of the following random
π‘Ž
π‘Œ
variables a) 𝐴 = ∏𝑛𝑖=1 π‘Œπ‘– , b) 𝐡 = ∏𝑛𝑖=1 π‘Œπ‘– , c) 𝐢 = π‘Œ1, d) find 𝐸(𝐴) = 𝐸(∏𝑛𝑖=1 π‘Œπ‘– ).
2
a) We have that 𝑙𝑛(𝐴) = 𝑙𝑛(∏𝑛𝑖=1 π‘Œπ‘– ) = ∑𝑛𝑖=1 𝑙𝑛⁑(π‘Œπ‘– ) = 𝑙𝑛(π‘Œ1 ) + β‹― + 𝑙𝑛⁑(π‘Œπ‘› ), so the
random variable 𝑙𝑛⁑(𝐴) is the sum of 𝑛 normally distributed random variables. This
implies that 𝑙𝑛(𝐴) ~𝑁(∑𝑛𝑖=1 πœ‡π‘– , ∑𝑛𝑖=1 πœŽπ‘–2 ), which means 𝐴~𝐿𝑂𝐺𝑁(∑𝑛𝑖=1 πœ‡π‘– , ∑𝑛𝑖=1 πœŽπ‘–2 ).
b) The
random
variable
𝑙𝑛(𝐡) = 𝑙𝑛(∏𝑛𝑖=1 π‘Œπ‘–π‘Ž ) = ∑𝑛𝑖=1 𝑙𝑛⁑(π‘Œπ‘–π‘Ž ) = ∑𝑛𝑖=1 π‘Žπ‘– 𝑙𝑛⁑(π‘Œπ‘– ) =
π‘Ž1 𝑙𝑛(π‘Œ1 ) + β‹― + π‘Ž2 𝑙𝑛⁑(π‘Œπ‘› ). We use that if some 𝑋~𝑁(πœ‡, 𝜎 2 ), then π‘Žπ‘‹~𝑁(π‘Žπœ‡, π‘Ž2 𝜎 2 ) to
conclude that 𝑙𝑛(𝐡) ~𝑁(∑𝑛𝑖=1 π‘Žπ‘– πœ‡π‘– , ∑𝑛𝑖=1 π‘Žπ‘–2 πœŽπ‘–2 ) so 𝐡~𝐿𝑂𝐺𝑁(∑𝑛𝑖=1 π‘Žπ‘– πœ‡π‘– , ∑𝑛𝑖=1 π‘Žπ‘–2 πœŽπ‘–2 ).
π‘Œ
c) We have that 𝑙𝑛(𝐢) = 𝑙𝑛 (π‘Œ1 ) = 𝑙𝑛(π‘Œ1 ) − 𝑙𝑛(π‘Œ2 ), so the random variable 𝑙𝑛⁑(𝐢) is the
2
sum of two normally distributed random variables. Thus, 𝑙𝑛(𝐢) ~𝑁(πœ‡1 − πœ‡2 , 𝜎12 + 𝜎22 )
π‘Œ
which implies that the distribution of 𝐢 = π‘Œ1 is 𝐢~𝐿𝑂𝐺𝑁(πœ‡1 − πœ‡2 , 𝜎12 + 𝜎22 ).
2
d) For 𝑋~𝑁(πœ‡, 𝜎 2 ), we have 𝑀𝑋 (𝑑) = 𝐸(𝑒 𝑑𝑋 ) = 𝑒 πœ‡π‘‘+𝜎
𝑀𝑍 (𝑑) = 𝐸(𝑒 𝑑𝑍 ) = 𝑒 𝑑
2 /2
2 𝑑 2 /2
and for 𝑍~𝑁(0,1), we have
. Thus, the expected value is given by 𝐸(π‘Œπ‘– ) = 𝐸(𝑒 πœ‡π‘– +πœŽπ‘– 𝑍 ) =
2
𝑒 πœ‡π‘– +πœŽπ‘– /2. Since the random variables π‘Œπ‘– are all independent, we therefore have that
2
𝐸(𝐴) = 𝐸(∏𝑛𝑖=1 π‘Œπ‘– ) = ∏𝑛𝑖=1 𝐸(π‘Œπ‘– ) = ∏𝑛𝑖=1(𝑒 πœ‡π‘– +πœŽπ‘– /2 ) = 𝑒π‘₯𝑝⁑{∑𝑛𝑖=1 πœ‡π‘– + ∑𝑛𝑖=1 πœŽπ‘–2 /2}.
Question #28: Let 𝑋1 and 𝑋2 be a random sample of size 2 from a continuous distribution
with PDF of the form 𝑓𝑋 (π‘₯) = 2π‘₯ if 0 < π‘₯ < 1 and zero otherwise. a) Find the marginal
densities of π‘Œ1 and π‘Œ2 , the smallest and largest order statistics, b) find the joint probability
density function of π‘Œ1 and π‘Œ2 , and c) find the density of the sample range 𝑅 = π‘Œ2 − π‘Œ1 .
0⁑⁑⁑⁑⁑⁑⁑𝑖𝑓⁑π‘₯ ∈ (−∞, 0]⁑
(0,1)⁑⁑
2π‘₯⁑⁑⁑⁑⁑𝑖𝑓π‘₯
∈
a) Since 𝑓𝑋 (π‘₯) = {
, we know that 𝐹𝑋 (π‘₯) = { π‘₯ 2 ⁑⁑⁑⁑⁑𝑖𝑓⁑π‘₯ ∈ (0,1)⁑⁑⁑⁑⁑⁑ . Then
0β‘β‘β‘β‘β‘β‘β‘β‘π‘œπ‘‘β„Žπ‘’π‘Ÿπ‘€π‘–π‘ π‘’β‘β‘β‘
1⁑⁑⁑⁑⁑⁑⁑⁑𝑖𝑓⁑π‘₯ ∈ [1, ∞)⁑⁑⁑⁑
from Theorem 6.5.2, we have that 𝑔1 (𝑦1 ) = 𝑛𝑓𝑋 (𝑦1 )⁑[1 − 𝐹𝑋 (𝑦1 )]𝑛−1 so we can
calculate the smallest order statistic as 𝑔1 (𝑦1 ) = 2[2𝑦1 ][1 − 𝑦12 ]2−1 = 4𝑦1 − 4𝑦13
whenever 𝑦1 ∈ (0,1). Similarly, 𝑔𝑛 (𝑦𝑛 ) = 𝑛𝑓𝑋 (𝑦𝑛 )⁑[𝐹𝑋 (𝑦𝑛 )]𝑛−1 so we can calculate the
largest order statistic as 𝑔2 (𝑦2 ) = 2[2𝑦2 ][𝑦22 ]2−1 = 4𝑦23 whenever 𝑦2 ∈ (0,1).
b) From Theorem 6.5.1, the joint probability density function of the order statistics is
𝑔𝒀 (𝑦1 , … , 𝑦𝑛 ) = 𝑛! 𝑓𝑋 (𝑦1 ) … 𝑓𝑋 (𝑦𝑛 ). In this question, we have that π‘”π‘Œ1 π‘Œ2 (𝑦1 , 𝑦2 ) =
2! 𝑓𝑋 (𝑦1 )𝑓𝑋 (𝑦2 ) = 2(2𝑦1 )(2𝑦2 ) = 8𝑦1 𝑦2 whenever we have 0 < 𝑦1 < 𝑦2 < 1.
c) We first find the joint density of the smallest and largest order statistics in order to
make a transformation to get the marginal density of the sample range. From the
work we did above, we have that π‘”π‘Œ1 π‘Œ2 (𝑦1 , 𝑦2 ) = 8𝑦1 𝑦2 . We have the transformation
π‘Ÿ = 𝑦2 − 𝑦1 and generate 𝑠 = 𝑦1 , so we have 𝑦1 = 𝑠 and 𝑦2 = π‘Ÿ + 𝑠 which allows us to
1 0
calculate 𝐽 = 𝑑𝑒𝑑 [
] = 1. The joint density of 𝑆 and 𝑅 is therefore 𝑓𝑆𝑅 (𝑠, π‘Ÿ) =
1 1
π‘”π‘Œ1 π‘Œ2 (𝑠, π‘Ÿ + 𝑠)|𝐽| = 8𝑠(π‘Ÿ + 𝑠)(1) = 8𝑠 2 + 8π‘ π‘Ÿ if 0 < 𝑠 < π‘Ÿ + 𝑠 < 1, which can also be
∞
written as 0 < 𝑠 < 1 − π‘Ÿ. The marginal density is thus 𝑓𝑅 (π‘Ÿ) = ∫−∞ 𝑓𝑆𝑅 (𝑠, π‘Ÿ) 𝑑𝑠 =
1−π‘Ÿ
∫0
8
1−π‘Ÿ
(8𝑠 2 + 8π‘ π‘Ÿβ‘) 𝑑𝑠 = [3 𝑠 3 + 4π‘Ÿπ‘  2 ]
0
8
= 3 (1 − π‘Ÿ)3 + 4π‘Ÿ(1 − π‘Ÿ)2 if 0 < π‘Ÿ < 1.
Question #31: Consider a random sample of size 𝑛 from an exponential distribution such
that 𝑋𝑖 ~𝐸𝑋𝑃(1). Give the density of a) the smallest order statistic denoted by π‘Œ1 , b) the
largest order statistic denoted by π‘Œπ‘› , c) the sample range of the order statistics 𝑅 = π‘Œπ‘› − π‘Œ1 .
a) Since 𝑓𝑋𝑖 (π‘₯𝑖 ) = {
0⁑⁑⁑⁑⁑⁑⁑⁑⁑⁑⁑⁑⁑⁑⁑⁑𝑖𝑓π‘₯ ∈ (−∞, 0]
𝑒 −π‘₯𝑖 ⁑⁑⁑⁑⁑𝑖𝑓π‘₯ ∈ (0, ∞)⁑⁑
, we have 𝐹𝑋𝑖 (π‘₯𝑖 ) = {
.
1 − 𝑒 −π‘₯𝑖 β‘β‘β‘β‘β‘π‘œπ‘‘β„Žπ‘’π‘Ÿπ‘€π‘–π‘ π‘’β‘β‘β‘β‘β‘β‘
0β‘β‘β‘β‘β‘β‘β‘β‘β‘π‘œπ‘‘β„Žπ‘’π‘Ÿπ‘€π‘–π‘ π‘’β‘β‘β‘β‘β‘β‘
Then we have that 𝑔1 (𝑦1 ) = 𝑛𝑓𝑋 (𝑦1 )[1 − 𝐹𝑋 (𝑦1 )]𝑛−1 = 𝑛𝑒 −𝑦1 [1 − (1 − 𝑒 −𝑦1 )]𝑛−1 =
𝑛𝑒 −𝑦1 (𝑒 −𝑦1 )𝑛−1 = 𝑛𝑒 −𝑛𝑦1 if 𝑦1 > 0 and zero otherwise.
b) Similarly, we have 𝑔𝑛 (𝑦𝑛 ) = 𝑛𝑓𝑋 (𝑦𝑛 )[𝐹𝑋 (𝑦𝑛 )]𝑛−1 = 𝑛𝑒 −𝑦𝑛 [1 − 𝑒 −𝑦𝑛 ]𝑛−1 for 𝑦𝑛 > 0.
c) Since the exponential distribution has the memoryless property, the difference of
𝑅 = π‘Œπ‘› − π‘Œ1 will not be conditional on the value of π‘Œ1 . This allows us to treat π‘Œ1 = 0,
so that 𝑅 = π‘Œπ‘› − π‘Œ1 = π‘Œπ‘› − 0 = π‘Œπ‘› . We then use the fact that the range of a set of 𝑛
order statistics from an exponential distribution is the same as the largest order
statistic from a set of 𝑛 − 1 order statistics. From above, we have that 𝑔𝑛 (𝑦𝑛 ) =
𝑛𝑒 −𝑦𝑛 [1 − 𝑒 −𝑦𝑛 ]𝑛−1 , so substituting 𝑛 − 1 gives 𝑔𝑅 (π‘Ÿ) = (π‘Ÿ − 1)[1 − 𝑒 −π‘Ÿ ]π‘Ÿ−2 𝑒 −π‘Ÿ .
Question #32: A system is composed of five independent components connected in series
one after the other. a) If the PDF of the time to failure of each component is 𝑋𝑖 ~𝐸𝑋𝑃(1), then
give the PDF of the time to failure of the system; b) if the components are connected in
parallel so that all must fail before the system fails, give the PDF of the time to failure.
a) Since 𝑋𝑖 ~𝐸𝑋𝑃(1), we know that the density is given by 𝑓𝑋𝑖 (π‘₯𝑖 ) = 𝑒 −π‘₯𝑖 for π‘₯𝑖 > 0 and
0⁑⁑⁑⁑⁑⁑⁑⁑⁑⁑⁑⁑⁑⁑⁑⁑𝑖𝑓π‘₯ ∈ (−∞, 0]
𝐹𝑋𝑖 (π‘₯𝑖 ) = {
. The system in the series fails whenever the
1 − 𝑒 −π‘₯𝑖 β‘β‘β‘β‘β‘π‘œπ‘‘β„Žπ‘’π‘Ÿπ‘€π‘–π‘ π‘’β‘β‘β‘β‘β‘β‘
earliest component fails, which happens at time 𝑋(1) = π‘Œ1 , the first order statistic.
Thus, the probability density function of the time to failure is therefore given by
π‘“π‘Œ1 (𝑦1 ) = 𝑛𝑓𝑋 (𝑦1 )[1 − 𝐹𝑋 (𝑦1 )]𝑛−1 = 5𝑒 −𝑦1 [𝑒 −𝑦1 ]4 = 5𝑒 −5𝑦1 whenever 𝑦1 > 0.
b) For the system in parallel, the system in the series fails whenever the last component
fails, which happens at time 𝑋(5) = π‘Œ5, the greatest order statistic. Thus, the density is
π‘“π‘Œ5 (𝑦5 ) = 𝑛𝑓𝑋 (𝑦𝑛 )[𝐹𝑋 (𝑦𝑛 )]𝑛−1 = 5𝑒 −𝑦5 [1 − 𝑒 −𝑦5 ]4 whenever 𝑦5 > 0.
Question #33: Consider a random sample of size 𝑛 from a geometric distribution such that
𝑋𝑖 ~𝐺𝐸𝑂(𝑝). Give the CDF of a) the minimum π‘Œ1 , b) the π‘˜ π‘‘β„Ž smallest π‘Œπ‘˜ , c) the maximum π‘Œπ‘› .
a) If some 𝑋𝑖 ~𝐺𝐸𝑂(𝑝), then 𝑓𝑋𝑖 (π‘₯) = 𝑝(1 − 𝑝)π‘₯ and 𝐹𝑋𝑖 (π‘₯) = 1 − (1 − 𝑝)π‘₯+1. Consider
the event 𝑋(1) > π‘š, which happens if and only if 𝑋(𝑖) > π‘š for all 𝑖 = 1, … , 𝑛. Therefore,
𝑛
𝑛
we have 𝑃(𝑋(1) > π‘š) = 𝑃(𝑋(𝑖) > π‘š) = [1 − 𝐹𝑋𝑖 (π‘š)] = [(1 − 𝑝)π‘š ]𝑛 = (1 − 𝑝)π‘šπ‘› ,
which implies that the CDF is given by 𝐹𝑋(1) (π‘š) = 𝑃(𝑋(1) ≤ π‘š) = 1 − (1 − 𝑝)π‘šπ‘› .
b) The event 𝑋(π‘˜) ≤ π‘š happens when π‘˜ of the 𝑋(𝑖) satisfy 𝑋(𝑖) ≤ π‘š and the other 𝑛 − π‘˜
𝑛
satisfy 𝑋(𝑖) > π‘š. Thus, we have 𝑃(𝑋(π‘˜) ≤ π‘š) = (π‘›π‘˜)𝑃(𝑋(𝑖) ≤ π‘š) 𝑃(𝑋(𝑖) > π‘š)
𝑛−π‘˜
=
(π‘›π‘˜)[1 − (1 − 𝑝)π‘š ]π‘˜ [(1 − 𝑝)π‘š ]𝑛−π‘˜ , which is the distribution function of π‘Œπ‘˜ .
c) The event 𝑋(𝑛) ≤ π‘š happens when 𝑋(𝑖) ≤ π‘š for all 𝑖 = 1, … , 𝑛. Thus, 𝑃(𝑋(𝑛) ≤ π‘š) =
𝑛
𝑛
𝑃(𝑋(𝑖) ≤ π‘š) = (∑𝑛𝑗=1[𝑝(1 − 𝑝)𝑗−1 ]) = (𝑝
1−(1−𝑝)π‘š 𝑛
1−(1−𝑝)
) = [1 − (1 − 𝑝)π‘š ]𝑛 .
Chapter #7 – Limiting Distributions
Question #30: If 𝑋~𝑃𝐴𝑅(πœƒ, πœ…), then 𝑓𝑋 (π‘₯) =
πœ…
π‘₯ πœ…+1
πœƒ
πœƒ(1+ )
if π‘₯ > 0 and zero otherwise. Consider
a random sample of size 𝑛 = 5 from a Pareto distribution where 𝑋~𝑃𝐴𝑅(1,2); that is,
suppose that 𝑋1 , … , 𝑋5 are drawn from the given Pareto distribution above. a) Find the joint
PDF of the second and fourth order statistics given by π‘Œ2 = 𝑋(2) and π‘Œ4 = 𝑋(4) , and b) find the
joint PDF of the first three order statistics given by π‘Œ1 = 𝑋(2) , π‘Œ2 = 𝑋(2) and π‘Œ3 = 𝑋(2).
π‘₯
π‘₯
2
a) The CDF of the population is given by 𝐹𝑋 (π‘₯) = ∫0 𝑓𝑋 (𝑑) 𝑑𝑑 = ∫0 [(1+𝑑)3 ] 𝑑𝑑, so that we
can calculate the joint density using Corollary 6.5.1 as
5!
[𝐹 (𝑦 )]2−1 𝑓𝑋 (𝑦2 )[𝐹𝑋 (𝑦4 )
(2−1)!(4−2−1)!(5−4)! 𝑋 2
π‘“π‘Œ2 π‘Œ4 (𝑦2 , 𝑦4 ) =
− 𝐹𝑋 (𝑦2 )]4−2−1 𝑓𝑋 (𝑦4 )[1 − 𝐹𝑋 (𝑦4 )]5−4 =
5! 𝑓𝑋 (𝑦2 )𝑓𝑋 (𝑦4 )[𝐹𝑋 (𝑦2 )][𝐹𝑋 (𝑦4 ) − 𝐹𝑋 (𝑦2 )][1 − 𝐹𝑋 (𝑦4 )] if 0 < 𝑦2 < 𝑦4 < ∞.
𝑛!
b) From Theorem 6.5.4, we have 𝑔(𝑦1 , … , π‘¦π‘Ÿ ) = (𝑛−π‘Ÿ)! [1 − 𝐹𝑋 (π‘¦π‘Ÿ )]𝑛−π‘Ÿ [𝑓𝑋 (𝑦1 ) … 𝑓𝑋 (π‘¦π‘Ÿ )],
so we may calculate that π‘“π‘Œ1 π‘Œ2 π‘Œ3 (𝑦1 , 𝑦2 , 𝑦3 ) = 60[1 − 𝐹𝑋 (𝑦3 )]2 [𝑓𝑋 (𝑦1 )𝑓𝑋 (𝑦2 )𝑓𝑋 (𝑦3 )]
𝑛!
5!
whenever 0 < 𝑦1 < 𝑦2 < 𝑦3 < ∞, since we have that (𝑛−π‘Ÿ)! = 2! = 60.
Question #1: Consider a random sample of size 𝑛 from a distribution with cumulative
1
distribution function 𝐹𝑋 (π‘₯) = 1 − π‘₯ whenever 1 ≤ π‘₯ < ∞ and zero otherwise. That is, let the
random variables 𝑋1 , … , 𝑋𝑛 be ~𝑖𝑖𝑑 from the distribution with CDF 𝐹𝑋 (π‘₯). a) Derive the CDF
of the smallest order statistic given by 𝑋(1) = 𝑋1:𝑛 , b) find the limiting distribution of 𝑋1:𝑛 ;
that is, if 𝐺𝑛 (𝑦) denotes the order statistic from above, find lim 𝐺𝑛 (𝑦), c) find the limiting
𝑛→∞
distribution of
𝑛
𝑋1:𝑛
;
that is, find the CDF of
𝑛
𝑋(1)
and its limit as 𝑛 → ∞.
a) We can compute that 𝐹𝑋1:𝑛 (𝑦) = 𝑃(𝑋1:𝑛 ≤ 𝑦) = 1 − 𝑃(𝑋1:𝑛 > 𝑦) = 1 − 𝑃(𝑋𝑖 > 𝑦)𝑛 =
1
𝑛
1 𝑛
1 − [1 − (1 − 𝑦)] = 1 − (𝑦) whenever 1 ≤ 𝑦 < ∞. We thus have that the CDF of
the smallest order statistic is 𝐹𝑋1:𝑛 (𝑦) = {
1 − (1/𝑦)𝑛
0
𝑖𝑓 𝑦 ≥ 1
. Finally, we note
𝑖𝑓 𝑦 < 1
that 𝑃(𝑋1:𝑛 > 𝑦) ≡ 𝑃(𝑋𝑖 > 𝑦)𝑛 since the smallest order statistic is greater than some
𝑦 if and only if all 𝑛 of the independent samples are also greater than 𝑦. We can use
this approach for any order statistic, including the largest, by changing the exponent.
b) We have that lim 𝐺𝑛 (𝑦) = {
𝑛→∞
lim [1 − (1/𝑦)𝑛 ]
𝑖𝑓 𝑦 ≥ 1
lim 0
𝑖𝑓 𝑦 < 1
𝑛→∞
𝑛→∞
={
0 𝑖𝑓 𝑦 ≤ 1
= 𝐺(𝑦),
1 𝑖𝑓 𝑦 > 1
so the limiting distribution of 𝑋1:𝑛 is degenerate at 𝑦 = 1. From Definition 7.2.2, this
means that 𝐺(𝑦) is the cumulative distribution function of some discrete distribution
𝑔(𝑦) that assigns probability one at 𝑦 = 1 and zero otherwise.
1
1
𝑛
𝑛 (𝑦) = 𝑃(𝑋1:𝑛 ≤ 𝑦) = 𝑃 (𝑋1:𝑛 ≤ 𝑦 𝑛 ) = 1 − 𝑃 (𝑋1:𝑛 > 𝑦 𝑛 ) =
c) As before, we have 𝐹𝑋1:𝑛
1
𝑛
𝑛
1 − 𝑃 (𝑋𝑖 > 𝑦 ) = 1 − [1 − (1 −
1
1
𝑦𝑛
𝑛
1
)] = 1 − (𝑦) whenever 𝑦 ≥ 1. Therefore, it is
clear that the limiting distribution of this sequence of random variables is given by
𝑛 (𝑦) = {
lim 𝐹𝑋1:𝑛
𝑛→∞
1 − 1/𝑦
0
𝑖𝑓 𝑦 ≥ 1
= 𝐺(𝑦) since there is no dependence on 𝑛.
𝑖𝑓 𝑦 < 1
Question #2: Consider a random sample of size 𝑛 from a distribution with CDF given by
1
𝐹(π‘₯) = 1+𝑒 −π‘₯ for all π‘₯ ∈ ℝ. Find the limiting distribution of a) 𝑋𝑛:𝑛 and b) 𝑋𝑛:𝑛 − 𝑙𝑛(𝑛).
1
𝑛
a) We have 𝐹𝑋𝑛:𝑛 (𝑦) = 𝑃(𝑋𝑛:𝑛 ≤ 𝑦) = 𝑃(𝑋𝑖 ≤ 𝑦)𝑛 = (1+𝑒 −𝑦 )
1
𝑛
for all 𝑦 ∈ ℝ. Since
lim [(1+𝑒 −𝑦 ) ] = 0, we conclude that 𝑋𝑛:𝑛 does not have a limiting distribution.
𝑛→∞
b) We calculate that 𝐹𝑋𝑛:𝑛−𝑙𝑛(𝑛) (𝑦) = 𝑃(𝑋𝑛:𝑛 − 𝑙𝑛(𝑛) ≤ 𝑦) = 𝑃(𝑋𝑛:𝑛 ≤ 𝑦 + 𝑙 𝑛(𝑛)) =
𝑛
1
𝑛
1
𝐹𝑋𝑛:𝑛 (𝑦 + 𝑙 𝑛(𝑛)) = (1+𝑒 −(𝑦+𝑙 𝑛(𝑛)) ) = (1+𝑒 −𝑦 𝑒 −𝑙𝑛(𝑛)) = (
1+
limit gives lim [𝐹𝑋𝑛:𝑛 (𝑦 + 𝑙 𝑛(𝑛))] = lim [(
𝑛→∞
1
𝑒−𝑦
1+
𝑛
𝑛→∞
𝑛
1
𝑒−𝑦
𝑛
) .
Evaluating
this
𝑛
) ] = 𝑒𝑒
−𝑦
for all 𝑦 ∈ ℝ.
Question #3: Consider a random sample of size 𝑛 from the distribution 𝐹(π‘₯) = 1 − π‘₯ −2 if
π‘₯ > 1 and zero otherwise. Find the limiting distribution of a) 𝑋1:𝑛 , b) 𝑋𝑛:𝑛 and c)
1
√𝑛
𝑋𝑛:𝑛 .
a) We can compute that 𝐹𝑋1:𝑛 (𝑦) = 𝑃(𝑋1:𝑛 ≤ 𝑦) = 1 − 𝑃(𝑋1:𝑛 > 𝑦) = 1 − 𝑃(𝑋𝑖 > 𝑦)𝑛 =
1
𝑛
1
1
1 − [1 − (1 − 𝑦 2 )] = 1 − (𝑦 2𝑛) if 𝑦 > 1. Thus, 𝐹𝑋1:𝑛 (𝑦) = {
1
lim [1 − 𝑦 2𝑛]
1 − 𝑦 2𝑛
𝑖𝑓 𝑦 > 1
0
𝑖𝑓 𝑦 ≤ 1
so
𝑖𝑓 𝑦 > 1
1 𝑖𝑓 𝑦 > 1
={
.
0 𝑖𝑓 𝑦 ≤ 1
𝑖𝑓 𝑦 ≤ 1
the limiting distribution is lim 𝐹𝑋1:𝑛 (𝑦) = {𝑛→∞
𝑛→∞
lim [0]
𝑛→∞
We therefore say that the limiting distribution is degenerate at 𝑦 = 1.
𝑛
1
b) We have 𝐹𝑋𝑛:𝑛 (𝑦) = 𝑃(𝑋𝑛:𝑛 ≤ 𝑦) = 𝑃(𝑋𝑖 ≤ 𝑦)𝑛 = (1 − 𝑦 2 ) whenever 𝑦 > 1. Thus,
1
(1 − 𝑦 2 )
𝐹𝑋𝑛:𝑛 (𝑦) = {
0
𝑛
1
𝑛
lim [(1 − 𝑦 2 ) ] 𝑖𝑓 𝑦 > 1
𝑖𝑓 𝑦 > 1
so lim 𝐹𝑋𝑛:𝑛 (𝑦) = {𝑛→∞
𝑛→∞
lim [0]
𝑖𝑓 𝑦 ≤ 1
𝑛→∞
𝑖𝑓 𝑦 ≤ 1
= 0.
We would therefore conclude that there is no limiting distribution for 𝑋𝑛:𝑛 .
c) We compute that 𝐹 1 𝑋
√𝑛
(1 − (
1
√𝑛𝑦)2
𝑛
1
𝑛:𝑛
(𝑦) = 𝑃 (
1
√𝑛
𝑋𝑛:𝑛 ≤ 𝑦) = 𝑃(𝑋𝑛:𝑛 ≤ √𝑛𝑦) = 𝐹𝑋𝑛:𝑛 (√𝑛𝑦) =
𝑛
1
) = (1 − 𝑛𝑦 2 ) whenever √𝑛𝑦 > 1 or 𝑦 > 𝑛. We can therefore compute
√
1
the limit as lim 𝐹 1 𝑋
𝑛→∞
√𝑛
𝑛:𝑛
(𝑦) = {
𝑛
lim [(1 − 𝑛𝑦 2 ) ] 𝑖𝑓 𝑦 >
𝑛→∞
lim [0]
𝑛→∞
𝑖𝑓 𝑦 ≤
1
√𝑛
1
√𝑛
1
− 2
𝑦
= {𝑒
0
𝑖𝑓 𝑦 > 0 .
𝑖𝑓 𝑦 ≤ 0
Question #5: Suppose that 𝑍𝑖 ~𝑁(0,1) and that the 𝑍𝑖 are all independent. Use moment
generating functions to find the limiting distribution of 𝐴𝑛 =
ο‚·
We have 𝐴𝑛 =
1
𝑛
∑𝑛
𝑖=1(𝑍𝑖 + )
√𝑛
1
𝑛
𝑛
(∑𝑛
𝑖=1 𝑍𝑖 )+(∑𝑖=1( ))
=
√𝑛
=
1
1
𝑛
∑𝑛
𝑖=1(𝑍𝑖 + )
√𝑛
(∑𝑛
𝑖=1 𝑍𝑖 )+𝑛
1
𝑛
√𝑛
=
as 𝑛 → ∞.
∑𝑛
𝑖=1 𝑍𝑖
√𝑛
+
1
√𝑛
, so the MGF is
𝑑
𝑀𝐴𝑛 (𝑑) = [𝑀1 (𝑑)]𝑛 [𝑀2 (𝑑)] = [𝑀1 (𝑑)]𝑛 [𝐸 (𝑒 √𝑛 )] since 𝐴𝑛 is the sum of two parts so
we can multiply their respective MGFs. The MGF of a standard normal random
variable with πœ‡ = 0 and 𝜎 2 = 1 is given by 𝑀𝑍 (𝑑) = 𝑒 𝑑
that 𝑀1 (𝑑) = 𝑒
(
𝑑 2
) /2
√𝑛
𝑑2
2𝑛
2 /2
1
, which allows us to calculate
𝑑
𝑑
= 𝑒 . Also, we have that 𝐸 (𝑒 √𝑛 ) = 𝑒 √𝑛 so combining these
𝑛
gives 𝑀𝐴𝑛 (𝑑) = [𝑀1 (𝑑)] [𝐸 (𝑒
1
𝑑
√𝑛
𝑑2
2𝑛
𝑛
𝑑2
𝑑
𝑑
)] = [𝑒 ] [𝑒 √𝑛 ] = [𝑒 2 ] [𝑒 √𝑛 ]. Then we can use
𝑑2
𝑑
𝑑2
Theorem 7.3.1 to calculate lim 𝑀𝐴𝑛 (𝑑) = lim [𝑒 2 ] [𝑒 √𝑛 ] = 𝑒 2 = 𝑀(𝑑), which we
𝑛→∞
𝑛→∞
know is the MGF of a standard normal, so the limiting distribution is 𝐴~𝑁(0,1). Note
that this is also a direct consequence of the Central Limit Theorem.
Question #9: Let 𝑋1 , 𝑋2 , … , 𝑋100 be a random sample of size 𝑛 = 100 from an exponential
distribution such that each 𝑋𝑖 ~𝐸𝑋𝑃(1) and let π‘Œ = 𝑋1 + 𝑋2 + β‹― + 𝑋100. a) Give a normal
π‘Œ
approximation for the probability 𝑃(π‘Œ > 110), and b) if 𝑋̅ = 100 is the sample mean, then
give a normal approximation to the probability 𝑃(1.1 < 𝑋̅ < 1.2).
a) Since each 𝑋𝑖 ~𝐸𝑋𝑃(1), we know that 𝐸(𝑋𝑖 ) = πœƒ = 1 while π‘‰π‘Žπ‘Ÿ(𝑋𝑖 ) = πœƒ 2 = 1. Due to
100
the independence of the 𝑋𝑖 s, we have that 𝐸(π‘Œ) = 𝐸(∑100
𝑖=1 𝑋𝑖 ) = ∑𝑖=1 𝐸(𝑋𝑖 ) = 100 and
100
π‘‰π‘Žπ‘Ÿ(π‘Œ) = π‘‰π‘Žπ‘Ÿ(∑100
𝑖=1 𝑋𝑖 ) = ∑𝑖=1 π‘‰π‘Žπ‘Ÿ(𝑋𝑖 ) = 100 so 𝑠𝑑(π‘Œ) = √100 = 10. We can
therefore calculate that 𝑃(π‘Œ > 110) = 1 − 𝑃(π‘Œ ≤ 110) = 1 − 𝑃(∑100
𝑖=1 𝑋𝑖 ≤ 110) =
1−𝑃(
∑100
𝑖=1 𝑋𝑖 −100
10
≤
110−100
10
) ≈ 1 − 𝑃(𝑍 ≤ 1) = 1 − Φ(1) = 1 − 0.8413 = 0.1587,
where 𝑍 denotes the standard normal distribution with πœ‡ = 0 and 𝜎 2 = 1.
𝑋̅ −πœ‡
b) We know that 𝑍𝑛 = 𝜎/
√𝑛
→𝑑 𝑁(0,1) by the Central Limit Theorem. We then have that
π‘Œ
1
π‘Œ
1
1
𝐸(𝑋̅) = 𝐸 (100) = 100 𝐸(π‘Œ) = 1 and π‘‰π‘Žπ‘Ÿ(𝑋̅) = π‘‰π‘Žπ‘Ÿ (100) = 10,000 π‘‰π‘Žπ‘Ÿ(π‘Œ) = 100 so
1
1.1−1
𝑋̅ −1
𝑠𝑑(π‘Œ) = 10, which allows us to find 𝑃(1.1 < 𝑋̅ < 1.2) = 𝑃 ( 1/10 < 1/10 <
1.2−1
1/10
)≈
𝑃(1 < 𝑍𝑛 < 2) = Φ(2) − Φ(1) = 0.9772 − 0.8413 = 0.1359. Here, we have used the
fact that πœ‡ = 1 and 𝜎 = 1 which come from the population distribution 𝑋𝑖 ~𝐸𝑋𝑃(1).
Question #11: Let 𝑋𝑖 ~π‘ˆπ‘πΌπΉ(0,1) where 𝑋1 , 𝑋2 , … , 𝑋20 are all independent. Find a normal
approximation for the probability 𝑃(∑20
𝑖=1 𝑋𝑖 ≤ 12).
ο‚·
Since each 𝑋𝑖 ~π‘ˆπ‘πΌπΉ(0,1), we know that 𝐸(𝑋𝑖 ) = 1/2 while π‘‰π‘Žπ‘Ÿ(𝑋𝑖 ) = 1/12. Due to
20
the independence of the 𝑋𝑖 s, we have that 𝐸(∑20
𝑖=1 𝑋𝑖 ) = ∑𝑖=1 𝐸(𝑋𝑖 ) = 10 and
20
20
π‘‰π‘Žπ‘Ÿ(∑20
𝑖=1 𝑋𝑖 ) = ∑𝑖=1 π‘‰π‘Žπ‘Ÿ(𝑋𝑖 ) = 5/3, so that 𝑠𝑑(∑𝑖=1 𝑋𝑖 ) = √5/3. This allows us to
find 𝑃(∑20
𝑖=1 𝑋𝑖 ≤ 12) = 𝑃 (
∑20
𝑖=1 𝑋𝑖 −10
√5/3
≤
12−10
√5/3
) ≈ 𝑃(𝑍 ≤ 1.55) = Φ(1.55) = 0.9394.
Chapter #8 – Statistics and Sampling Distributions
Question #1: Let 𝑋 denote the weight in pounds of a single bag of feed where 𝑋~𝑁(101,4).
What is the probability that 20 bags will weigh at least 2,000 pounds?
ο‚·
20
20
Let π‘Œ = ∑20
𝑖=1 𝑋𝑖 where 𝑋𝑖 ~𝑁(101,4). We have that 𝐸(π‘Œ) = 𝐸(∑𝑖=1 𝑋𝑖 ) = ∑𝑖=1 𝐸(𝑋𝑖 ) =
20
20(101) = 2,020 and π‘‰π‘Žπ‘Ÿ(π‘Œ) = π‘‰π‘Žπ‘Ÿ(∑20
𝑖=1 𝑋𝑖 ) = ∑𝑖=1 π‘‰π‘Žπ‘Ÿ(𝑋𝑖 ) = 20(4) = 80 such
that 𝑠𝑑(π‘Œ) = √80 = 4√5. We can thus calculate the probability 𝑃(π‘Œ ≥ 2,000) =
∑20
𝑖=1 𝑋𝑖 −𝐸(π‘Œ)
𝑃(∑20
𝑖=1 𝑋𝑖 ≥ 2,000) = 𝑃 (
𝑠𝑑(π‘Œ)
≥
2,000−𝐸(π‘Œ)
∑20
𝑖=1 𝑋𝑖 −2,020
𝑠𝑑(π‘Œ)
4√5
) = 𝑃(
≥
2,000−2,020
)≈
4√5
20
𝑃 (𝑍 ≥ − 4√5) = 𝑃(𝑍 ≥ −2.24) = 1 − Φ(−2.24) = 0.987, where 𝑍~𝑁(0,1).
Question #2: Let 𝑆 denote the diameter of a shaft and 𝐡 the diameter of a bearing, where
both 𝑆 and 𝐡 are independent and 𝑆~𝑁(1,0.0004) and 𝐡~𝑁(1.01,0.0009). a) If a shaft and
bearing are selected at random, what is the probability that the shaft diameter will exceed
the bearing diameter? b) Now assume equal variances (πœŽπ‘†2 = 𝜎𝐡2 = 𝜎 2 ) such that we have
𝑆~𝑁(1, 𝜎 2 ) and 𝐡~𝑁(1.01, 𝜎 2 ). Find the value of 𝜎 that will yield a probability of
noninterference of 0.95 (which means the shaft diameter exceeds the bearing diameter).
a) Define π‘Œ = 𝑆 − 𝐷, since we wish to find 𝑃(𝑆 > 𝐷) = 𝑃(𝑆 − 𝐷 > 0) = 𝑃(π‘Œ > 0). We
have that 𝐸(π‘Œ) = 𝐸(𝑆 − 𝐷) = 𝐸(𝑆) − 𝐸(𝐷) = 1 − 1.01 = −0.01 and π‘‰π‘Žπ‘Ÿ(π‘Œ) =
π‘‰π‘Žπ‘Ÿ(𝑆 − 𝐷) = π‘‰π‘Žπ‘Ÿ(𝑆) + π‘‰π‘Žπ‘Ÿ(𝐷) = 0.0004 + 0.0009 = 0.0013 such that 𝑠𝑑(π‘Œ) =
√0.0013 = 0.036.
π‘Œ−𝐸(π‘Œ)
𝑃(
𝑠𝑑(π‘Œ)
>
0−𝐸(π‘Œ)
𝑠𝑑(π‘Œ)
Thus,
we
π‘Œ+0.01
have
𝑃(𝑆 > 𝐷) = 𝑃(𝑆 − 𝐷 > 0) = 𝑃(π‘Œ > 0) =
0.01
) = 𝑃 ( 0.036 > 0.036) ≈ 𝑃(𝑍 > 0.28) = 1 − Φ(0.28) = 0.39.
b) For π‘Œ = 𝑆 − 𝐷, we have that 𝐸(π‘Œ) = −0.01 but π‘‰π‘Žπ‘Ÿ(π‘Œ) = 2𝜎 2 so 𝑠𝑑(π‘Œ) = √2𝜎. We
wish to find 𝜎 so that 𝑃(π‘Œ > 0) = 0.95 → 1 − 𝑃 (𝑍 ≤
0.01
0.01
√2𝜎
√2𝜎
) = 0.95 → Φ (
) = 0.05.
Since only the critical value 𝑍𝛼 = −1.645 ensures that Φ(−1.645) = 0.05, we must
solve
0.01
√2𝜎
= −1.645 → 𝜎 = −0.004. But since we must have 𝜎 ≥ 0, no such 𝜎 exists.
Question #3: Let 𝑋1 , … , 𝑋𝑛 be a random sample of size 𝑛 where they are ~𝑖𝑖𝑑 such that
𝑋𝑖 ~𝑁(πœ‡, 𝜎 2 ) and define π‘ˆ = ∑𝑛𝑖=1 𝑋𝑖 and π‘Š = ∑𝑛𝑖=1 𝑋𝑖2 . a) Find a statistic that is a function of
π‘ˆ and π‘Š and unbiased for the parameter πœƒ = 2πœ‡ − 5𝜎 2 . b) Find a statistic that is unbiased
for 𝛾 = 𝜎 2 + πœ‡ 2 . c) If 𝑐 is a constant and π‘Œπ‘– = 1 if 𝑋𝑖 ≤ 𝑐 and zero otherwise, find a statistic
𝑐−πœ‡
that is a function of π‘Œ1 , … , π‘Œπ‘› and is unbiased for 𝐹𝑋 (𝑐) = Φ (
𝜎
2
)=
1
𝑑
(𝑐−πœ‡)/𝜎 1
𝑒− 2
∫−∞
√2πœ‹
1
𝑑𝑑.
1
a) We first find an estimator for πœ‡ = 𝐸(𝑋̅) = 𝐸 (𝑛 ∑𝑛𝑖=1 𝑋𝑖 ) = 𝑛 𝐸(∑𝑛𝑖=1 𝑋𝑖 ) = 𝑛 𝐸(π‘ˆ) =
1
and then for 𝜎 2 = 𝐸(𝑆 2 ) = 𝐸 (
𝑛−1
1
∑𝑛𝑖=1(𝑋𝑖 − 𝑋̅)2 ) = 𝐸 (
1
1
𝑛−1
π‘ˆ
𝑛
[∑𝑛𝑖=1(𝑋𝑖2 ) − 𝑛𝑋̅ 2 ]) =
1
2
1
𝐸(∑𝑛𝑖=1(𝑋𝑖2 ) − 𝑛𝑋̅ 2 ) = 𝑛−1 𝐸(∑𝑛𝑖=1 𝑋𝑖2 − 𝑛𝑋̅ 2 ) = 𝑛−1 𝐸 [π‘Š − 𝑛 (𝑛 ∑𝑛𝑖=1 𝑋𝑖 ) ] =
𝑛−1
1
𝐸 [π‘Š − 𝑛
𝑛−1
(∑𝑛
𝑖=1 𝑋𝑖 )
𝑛2
2
1
] = 𝑛−1 𝐸 [π‘Š −
(∑𝑛
𝑖=1 𝑋𝑖 )
2
𝑛
1
1
π‘ˆ
1
thus have πœƒ = 2πœ‡ − 5𝜎 2 = 2 [ 𝑛 ] − 5 [𝑛−1 (π‘Š −
We
π‘ˆ2
1
] = 𝑛−1 𝐸 [π‘Š − 𝑛 π‘ˆ 2 ] = 𝑛−1 [π‘Š −
π‘ˆ2
𝑛
2π‘ˆ
)] =
𝑛
𝑛
5
].
1
− 𝑛−1 (π‘Š − 𝑛 π‘ˆ 2 ),
which is an unbiased estimator of πœƒ since 𝐸(πœƒ) = β‹― = πœƒ.
2
1
1
b) Since we found that πœ‡ = 𝐸(𝑋̅) = 𝐸 ( ∑𝑛𝑖=1 𝑋𝑖 ), then we have πœ‡ 2 = [𝐸 ( ∑𝑛𝑖=1 𝑋𝑖 )] =
𝑛
2
1
𝑛
1
1
1
1
𝐸 [(𝑛 ∑𝑛𝑖=1 𝑋𝑖 ) − π‘‰π‘Žπ‘Ÿ (𝑛 ∑𝑛𝑖=1 𝑋𝑖 )] = 𝐸 [𝑛2 (∑𝑛𝑖=1 𝑋𝑖 )2 − 𝑛2 π‘‰π‘Žπ‘Ÿ(∑𝑛𝑖=1 𝑋𝑖 )] = 𝐸 [𝑛2 π‘ˆ 2 −
π‘ˆ2
1
π‘›πœŽ 2 ] = 𝐸 [ 𝑛2 −
𝑛2
combining
(1−𝑛)
𝑛
1
𝜎2
]=
𝑛
these
[𝑛−1 (π‘Š −
π‘ˆ2
𝑛
π‘ˆ2
−
𝑛2
we
π‘ˆ2
𝜎2
find
1
1
. We previously found that 𝜎 2 = 𝑛−1 [π‘Š −
𝑛
π‘ˆ2
)] + 𝑛2 = 𝑛 (π‘Š −
π‘ˆ2
𝑛
𝜎2
𝛾 = 𝜎 2 + πœ‡ 2 = 𝜎 2 + 𝑛2 −
that
π‘ˆ2
) + 𝑛2 =
π‘Š
𝑛
π‘ˆ2
π‘ˆ2
𝑛
π‘Š
− 𝑛2 + 𝑛2 = 𝑛 ,
=
(1−𝑛)
𝑛
π‘ˆ2
𝑛
], so
π‘ˆ2
𝜎 2 + 𝑛2 =
which
is
an
unbiased estimator of 𝛾 since 𝐸(𝛾) = β‹― = 𝛾.
𝑋𝑖 −πœ‡
c) We have 𝑃(π‘Œπ‘– = 1) = 𝑃(𝑋𝑖 ≤ 𝑐) = 𝑃 (
and
𝜎
≤
𝑐−πœ‡
𝜎
) = 𝑃 (𝑍 ≤
𝑐−πœ‡
𝜎
𝑐−πœ‡
𝐸(π‘Œπ‘– ) = 1 βˆ™ 𝑃(π‘Œπ‘– = 1) + 0 βˆ™ 𝑃(π‘Œπ‘– = 0) = 𝑃(π‘Œπ‘– = 1) = Φ (
𝑐−πœ‡
) = Φ(
𝜎
𝜎
) = 𝐹𝑋 (𝑐)
) = 𝐹𝑋 (𝑐).
Then,
1
1
1
1
𝑐−πœ‡
𝑐−πœ‡
𝐸(π‘ŒΜ…) = 𝐸 (𝑛 ∑𝑛𝑖=1 π‘Œπ‘– ) = 𝑛 𝐸(∑𝑛𝑖=1 π‘Œπ‘– ) = 𝑛 ∑𝑛𝑖=1 𝐸(π‘Œπ‘– ) = 𝑛 π‘›Φ ( 𝜎 ) = Φ ( 𝜎 ) = 𝐹𝑋 (𝑐),
which means that π‘ŒΜ… is an unbiased estimator of 𝐹𝑋 (𝑐) = Φ((𝑐 − πœ‡)/𝜎).
Question #4: Assume that 𝑋1 and 𝑋2 are independent normal random variables such that
each 𝑋𝑖 ~𝑁(πœ‡, 𝜎 2 ) and define π‘Œ1 = 𝑋1 + 𝑋2 and π‘Œ2 = 𝑋1 − 𝑋2. Show that the random variables
π‘Œ1 and π‘Œ2 are independent and normally distributed.
ο‚·
Since 𝑋1 and 𝑋2 are independent normal random variables, we know that their joint
1
density function is 𝑓𝑋1 𝑋2 (π‘₯1 , π‘₯2 ) = 𝑓𝑋1 (π‘₯1 )𝑓𝑋2 (π‘₯2 ) = [
√2πœ‹πœŽ
1
𝑒
2πœ‹πœŽ2
−
1
[(π‘₯1 −πœ‡)2 +(π‘₯2 −πœ‡)2 ]
2𝜎2
π‘“π‘Œ1 π‘Œ2 (𝑦1 , 𝑦2 ) = 𝑓𝑋1 𝑋2 (
(π‘₯ −πœ‡)2
(π‘₯ −πœ‡)2
1
− 1 2
− 2
2𝜎
][
𝑒 2𝜎2 ]
√2πœ‹πœŽ
=
. We have the transformation 𝑦1 = π‘₯1 + π‘₯2 and 𝑦2 = π‘₯1 − π‘₯2 ,
which can be solved to obtain π‘₯1 =
the Jacobian 𝐽 = 𝑑𝑒𝑑 [
𝑒
1/2
1/2
,
2
and π‘₯2 =
𝑦1 −𝑦2
2
. This allows us to calculate
1/2
1
] = − 2, so we can compute the joint density
−1/2
𝑦1 +𝑦2 𝑦1 −𝑦2
2
𝑦1 +𝑦2
2
1
1
) |𝐽| = 2 [2πœ‹πœŽ2 𝑒
−
2
2
1
𝑦 +𝑦
𝑦 −𝑦
(( 1 2 −πœ‡) +( 1 2 −πœ‡) )
2
2
2𝜎2
1
simplifying this expression, we have π‘“π‘Œ1 π‘Œ2 (𝑦1 , 𝑦2 ) = 4πœ‹πœŽ2 𝑒
−
1
[𝑦 −2πœ‡]2
4𝜎2 1
𝑒
−
].
After
1
[𝑦 ]2
4𝜎2 2
. Since
the marginal densities can be separated, this shows that π‘Œ1 and π‘Œ2 are independent
and normally distributed. Moreover, we see that π‘Œ1 ~𝑁(2πœ‡, 2𝜎 2 ) and π‘Œ2 ~𝑁(0,2𝜎 2 ).
Question #12: The distance in feet by which a parachutist misses a target is 𝐷 = √𝑋12 + 𝑋22
where 𝑋1 and 𝑋2 are independent with each 𝑋𝑖 ~𝑁(0,25). Find the probability 𝑃(𝐷 ≤ 12.25).
ο‚·
We wish to find 𝑃(𝐷 ≤ 12.25) = 𝑃 (√𝑋12 + 𝑋22 ≤ 12.25) = 𝑃[𝑋12 + 𝑋22 ≤ (12.25)2 ] =
𝑃[(𝑋1 − 0)2 + (𝑋2 − 0)2 ≤ (12.25)2 ] = 𝑃[(𝑋1 − πœ‡)2 + (𝑋2 − πœ‡)2 ≤ (12.25)2 ] =
𝑃[
(𝑋1 −πœ‡)2
𝜎2
+
(𝑋2 −πœ‡)2
𝜎2
≤
(12.25)2
𝜎2
] = 𝑃 [∑2𝑖=1
(𝑋𝑖 −πœ‡)2
𝜎2
Since πœ‡ = 0 and 𝜎 2 = 25, we have 𝑃 [∑2𝑖=1
≤
(12.25)2
(𝑋𝑖 −0)2
25
𝜎2
≤
] = 𝑃 [∑2𝑖=1
(12.25)2
25
(𝑋𝑖 −0)2
𝜎2
≤
] ≈ 𝑃 [πœ’ 2 (2) ≤
(12.25)2
𝜎2
(12.25)2
25
].
]≈
𝑃[πœ’ 2 (2) ≤ 6] = 0.95. Note that we have used Corollary 8.3.4 to transform the
𝑋𝑖 −πœ‡ 2
question into one using the chi-square distribution, since ∑𝑛𝑖=1 (
because 𝑋𝑖 ~𝑁(πœ‡, 𝜎 2 ) implies that
𝑋𝑖 −πœ‡
𝜎
𝜎
) ~πœ’ 2 (𝑛). This is
𝑋𝑖 −πœ‡ 2
~𝑁(0,1) so (
𝜎
) ~πœ’ 2 (1) and that the sum of
𝑛 independent chi-square distributed random variables is distributed πœ’ 2 (𝑛).
Chapter #8 – Statistics and Sampling Distributions
Question #8: Suppose that 𝑋 and π‘Œ are independent and distributed 𝑋~πœ’ 2 (π‘š) and π‘Œ~πœ’ 2 (𝑛).
Is the random variable 𝑍 = π‘Œ − 𝑋 distributed chi-square if we have 𝑛 > π‘š?
ο‚·
No. The random variable 𝑍 = π‘Œ − 𝑋 can clearly take on negative values, whereas as a
random variable following the chi-square distribution must be positive.
Question #9: Suppose that 𝑋~πœ’ 2 (π‘š), 𝑆 = 𝑋 + π‘Œ~πœ’ 2 (π‘š + 𝑛) and that 𝑋 and π‘Œ are
independent random variables. Use moment generating functions to show that 𝑆 − 𝑋~πœ’ 2 (𝑛).
ο‚·
We know that if some 𝐴~πœ’ 2 (𝑣), then its MGF is given by π‘€π‘Œ (𝑑) = (1 − 2𝑑)−𝑣/2. We
thus have 𝑀𝑋 (𝑑) = (1 − 2𝑑)−π‘š/2 and 𝑀𝑆 (𝑑) = 𝑀𝑋+π‘Œ (𝑑) = (1 − 2𝑑)−(π‘š+𝑛)/2 . Since 𝑋
and π‘Œ are independent, we know that 𝑀𝑋+π‘Œ (𝑑) = 𝑀𝑋 (𝑑)π‘€π‘Œ (𝑑), which implies that
𝑀𝑆−𝑋 (𝑑) = 𝑀(𝑋+π‘Œ)−𝑋 (𝑑) = π‘€π‘Œ (𝑑) =
𝑀𝑋+π‘Œ (𝑑)
𝑀𝑋 (𝑑)
=
(1−2𝑑)−(π‘š+𝑛)/2
(1−2𝑑)−π‘š/2
= (1 − 2𝑑)−𝑛/2. Thus, we
have that π‘Œ = 𝑆 − 𝑋 is distributed chi-square with 𝑛 degrees of freedom.
Question #14: If 𝑇~𝑑(𝜈), find the distribution of the random variable 𝑇 2 .
ο‚·
We know that if 𝑍~𝑁(0,1) and 𝑉~πœ’ 2 (𝜈) are independent random variables, then the
distribution of 𝑇 =
𝑍2
produce 𝑇 2 = 𝑉/𝜈 =
𝑍
is Student’s t distribution. But then we can square this to
√𝑉/𝜈
𝑍 2 /1
𝑉/𝜈
, which makes it clear that 𝑇 2 ~𝐹(1, 𝜈). The reason for this is
that we know if some 𝑍~𝑁(0,1), then 𝑍 2 ~πœ’ 2 (1). Moreover, we are already given that
𝑉~πœ’ 2 (𝜈). Combining these results with the fact that if some 𝑉1 ~πœ’ 2 (𝜈1 ) and 𝑉2 ~πœ’ 2 (𝜈2 )
𝑉 /𝜈
are independent, then the random variable 𝑋 = 𝑉1 /𝜈1 ~𝐹(𝜈1 , 𝜈2 ). Therefore, 𝑇 2 follows
2
2
the F distribution with 1 and 𝜈 degrees of freedom whenever 𝑇~𝑑(𝜈).
Question #15: Suppose that 𝑋𝑖 ~𝑁(πœ‡, 𝜎 2 ) for 𝑖 = 1, … , 𝑛 and 𝑍𝑖 ~𝑁(0,1) for 𝑖 = 1, . . , π‘˜ and
that all variables are independent. Find the distribution of the following random variables.
a) 𝑋1 − 𝑋2 ~𝑁(πœ‡ − πœ‡, 𝜎 2 + 𝜎 2 ) ≡ 𝑁(0,2𝜎 2 )
b) 𝑋2 + 2𝑋3 ~𝑁(πœ‡ + 2πœ‡, 𝜎 2 + 4𝜎 2 ) ≡ 𝑁(3πœ‡, 5𝜎 2 )
c) 𝑍12 ~πœ’ 2 (1) since the square of a standard normal random variable is chi-square.
d)
𝑋1 −𝑋2
πœŽπ‘†π‘§ √2
~𝑑(π‘˜ − 1) since 𝑋1 − 𝑋2 ~𝑁(0,2𝜎 2 ) implies that
𝑋1 −𝑋2
𝜎√2
~𝑁(0,1) and dividing this
by the sample standard deviation 𝑆𝑧 of the Z sample makes it clear that ~𝑑(π‘˜ − 1).
e)
√𝑛(𝑋̅ −πœ‡)
~𝑑(π‘˜
πœŽπ‘†π‘§
√𝑛(𝑋̅ −πœ‡)
πœŽπ‘†π‘§
𝑋̅ −πœ‡
(π‘˜−1)𝑆𝑧2
√
𝜎2
− 1) since 𝑍 = 𝜎/
𝑍
𝑍
𝑧
√𝑉𝑧 /(π‘˜−1)
=𝑆 =
~𝑁(0,1), 𝑉𝑧 =
𝑛
~πœ’ 2 (π‘˜ − 1) and we can write
~𝑑(π‘˜ − 1) by the definition of the t distribution (see above).
f) 𝑍12 + 𝑍22 = πœ’ 2 (1) + πœ’ 2 (1)~πœ’ 2 (1 + 1) ≡ πœ’ 2 (2) since we can simply add the
parameters for a sum of independent chi-square random variables.
g) 𝑍12 − 𝑍22 → the distribution is unknown.
h)
i)
j)
𝑍1
√𝑍22
𝑍12
𝑍22
𝑍1
𝑍2
~𝑑(1) since 𝑉 = 𝑍22 ~πœ’ 2 (1) and we can write
𝑍1
√𝑍22
=
𝑍1
√𝑍22 /1
=
~𝐹(1,1) since 𝑉1 = 𝑍12 ~πœ’ 2 (1), 𝑉2 = 𝑍22 ~πœ’ 2 (1) and we have
𝑍1
√𝑉/1
𝑍12
𝑍22
=
~𝑑(1).
𝑉1 /1
= ~𝐹(1,1).
𝑉2 /1
𝑧
~πΆπ΄π‘ˆ(1,0) since we can generate the joint transformation 𝑒 = 𝑧1 and 𝑣 = 𝑧2 ,
2
1
calculate the joint density π‘“π‘ˆπ‘‰ (𝑒, 𝑣) and integrate out 𝑑𝑣 to find π‘“π‘ˆ (𝑒) = πœ‹(𝑒2 +1).
k)
𝑋̅
𝑍̅
→ the distribution is unknown.
l)
√π‘›π‘˜(𝑋̅ −πœ‡)
2
𝜎√∑π‘˜
𝑖=1 𝑍𝑖
𝑋̅ −πœ‡
~𝑑(π‘˜) since π‘Š = 𝜎/
expression
√π‘›π‘˜(𝑋̅ −πœ‡)
2
𝜎√∑π‘˜
𝑖=1 𝑍𝑖
=
√𝑛
Μ… −πœ‡)
√𝑛(𝑋
𝜎
π‘˜
2
~𝑁(0,1) and 𝑉 = ∑π‘˜π‘–=1 𝑍𝑖2 ~πœ’ 2 (π‘˜) and we can write the
=
√∑𝑖=1 𝑍𝑖
π‘Š
√𝑉/π‘˜
~𝑑(π‘˜) by the definition of the distribution.
π‘˜
m) ∑𝑛𝑖=1
(𝑋𝑖 −πœ‡)2
𝜎2
+ ∑π‘˜π‘–=1(𝑍𝑖 − 𝑍̅)2 ~πœ’ 2 (𝑛 + π‘˜ − 1) since ∑𝑛𝑖=1
(𝑋𝑖 −πœ‡)2
𝜎2
~πœ’ 2 (𝑛) by Corollary
2
(π‘˜−1)𝑆
8.3.4 and ∑π‘˜π‘–=1(𝑍𝑖 − 𝑍̅)2 = (π‘˜ − 1)𝑆𝑧2 = 12 𝑧 ~πœ’ 2 (π‘˜ − 1) by Theorem 8.3.6. Thus,
we have the sum of two chi-square random variables so we sum the parameters.
n)
𝑋̅
𝜎2
1
𝜎2
1
π‘˜
2
1
πœ‡
1
1
𝜎
+ π‘˜ ∑π‘˜π‘–=1 𝑍𝑖 ~𝑁 (𝜎2 , π‘›πœŽ2 + π‘˜) since 𝑋̅~𝑁 (πœ‡, 𝑛 ) implies that the random variable
2
1 𝑋
1
1
πœ‡
1
𝑋̅ = ∑𝑛𝑖=1 (𝜎2 ) 𝑛𝑖 ~𝑁 (∑π‘›π‘˜=1 (π‘›πœŽ2 ) πœ‡, ∑π‘›π‘˜=1 (π‘›πœŽ2 ) 𝜎 2 ) ≡ 𝑁 (𝜎2 , π‘›πœŽ2 ). Also, we have
1
∑π‘˜π‘–=1 𝑍𝑖 = 𝑍̅~𝑁 (0, ) so the distribution of their sum is normal and we sum their
π‘˜
𝑋̅
1
πœ‡
1
1
respective means and variances to conclude that 𝜎2 + π‘˜ ∑π‘˜π‘–=1 𝑍𝑖 ~𝑁 (𝜎2 , π‘›πœŽ2 + π‘˜).
2
o) π‘˜π‘Μ… 2 ~πœ’ 2 (1) since √π‘˜π‘Μ…~𝑁(0,1), so it must be that (√π‘˜π‘Μ…) = π‘˜π‘Μ… 2 ~πœ’ 2 (1).
p)
Μ… 2
(π‘˜−1) ∑𝑛
𝑖=1(𝑋𝑖 −𝑋)
~𝐹(𝑛
π‘˜
(𝑛−1)𝜎2 ∑𝑖=1(𝑍𝑖 −𝑍̅)2
Μ… 2
(π‘˜−1) ∑𝑛
𝑖=1(𝑋𝑖 −𝑋)
π‘˜
2
(𝑛−1)𝜎 ∑𝑖=1(𝑍𝑖 −𝑍̅)2
=
− 1, π‘˜ − 1) since we can simplify the random variable as
Μ… 2
12 ∑𝑛
𝑖=1(𝑋𝑖 −𝑋)
𝑛−1
Μ… 2
𝜎2 ∑π‘˜
𝑖=1(𝑍𝑖 −𝑍)
π‘˜−1
12 𝑆 2
= 𝜎2 𝑆𝑋2 and 12 𝑆𝑋2 ~πœ’ 2 (𝑛 − 1) and 𝜎 2 𝑆𝑍2 ~πœ’ 2 (π‘˜ − 1). We
𝑍
thus have the ratio of two chi-square random variables over their respective degrees
of freedom, which we know follows the F distribution.
Question #18: Assume that 𝑍~𝑁(0,1), 𝑉1 ~πœ’ 2 (5) and 𝑉2 ~πœ’ 2 (9) are all independent. Then
compute the probability that a) 𝑃(𝑉1 + 𝑉2 < 8.6), b) 𝑃 (
𝑍
√𝑉1 /5
𝑉
< 2.015), c) 𝑃(𝑍 > 0.611√𝑉2 ),
𝑉
1
d) 𝑃 (𝑉1 < 1.45) and e) find the value of 𝑏 such that 𝑃 (𝑉 +𝑉
< 𝑏) = 0.9.
2
1
2
a) Since 𝑉1 ~πœ’ 2 (5) and 𝑉2 ~πœ’ 2 (9), we know that 𝑉1 + 𝑉2 ~πœ’ 2 (14). This allows us to
compute 𝑃(𝑉1 + 𝑉2 < 8.6) = 0.144 using the tables for the chi-square distribution.
b) We know that if 𝑍~𝑁(0,1) and some 𝑉~πœ’ 2 (𝜈) are independent random variables,
then 𝑇 =
𝑍
𝑇=
√𝑉1 /5
𝑍
√𝑉/𝜈
follows the t distribution with 𝜈 degrees of freedom. We thus have that
𝑍
~𝑑(5), so we can compute 𝑃 (
√𝑉1 /5
< 2.015) = 0.95 using the t-table.
c) We wish to compute 𝑃(𝑍 > 0.611√𝑉2 ) = 𝑃 (
𝑍
√𝑉2
𝑃(
𝑍
√𝑉2 /9
> 0.611) = 𝑃 (
𝑍
3 > 0.611(3)) =
√𝑉2
> 1.833) = 0.05, from using the t-table since we know that
𝑉
𝑉 /5
9
𝑍
√𝑉2 /9
~𝑑(9).
𝑉 /5
d) We wish to compute 𝑃 (𝑉1 < 1.45) = 𝑃 (𝑉1 /9 < 1.45 (5)) = 𝑃 (𝑉1 /9 < 2.61). We know
2
2
2
that if some 𝑉1 ~πœ’ 2 (𝜈1 ) and 𝑉2 ~πœ’ 2 (𝜈2 ) are independent, then the random variable
𝑉 /𝜈
𝑉 /5
given by 𝑋 = 𝑉1 /𝜈1 ~𝐹(𝜈1 , 𝜈2 ). We thus have that 𝑉1 /9 ~𝐹(5,9), so we can therefore use
2
2
2
𝑉 /5
the F-table to compute the desired probability as 𝑃 (𝑉1 /9 < 2.61) = 0.9.
2
𝑉
𝑉1 +𝑉2
1
e) We wish to compute 𝑏 sich that 𝑃 (𝑉 +𝑉
< 𝑏) = 𝑃 (
1
𝑉
1
𝑉 /9
2
𝑉1
1
𝑉
1
> 𝑏) = 𝑃 (1 + 𝑉2 > 𝑏) =
1
5 1
𝑉 /9
𝑃 (𝑉2 > 𝑏 − 1) = 𝑃 (𝑉2/5 > 9 (𝑏 − 1)) = 0.9. But we know that 𝐹 = 𝑉2/5 ~𝐹(9,5) so we
1
1
1
can use tables to find that 𝑃(𝐹 > 0.383) = 0.9. This means that we must solve the
5 1
1
9
equation 9 (𝑏 − 1) = 0.383 → 𝑏 = 5 (0.383) + 1 → 𝑏 = 9
5
1
(0.383)+1
= 0.592.
1
1
Question #19: Suppose that 𝑇~𝑑(1). a) Show that the CDF of 𝑇 is 𝐹𝑇 (𝑑) = 2 + πœ‹ π‘Žπ‘Ÿπ‘π‘‘π‘Žπ‘›β‘(𝑑)
1
and b) show that the 100 ⋅ 𝛾 π‘‘β„Ž percentile is given by 𝑑𝛾 (1) = π‘‘π‘Žπ‘› [πœ‹ (𝛾 − 2)].
a) If some 𝑇~𝑑(𝜈), then its density is given by 𝑓𝑇 (𝑑) =
we have 𝑓𝑇 (𝑑) =
Γ(1)
1
Γ( )√πœ‹
2
1
1
𝜈+1
)
2
𝜈
Γ( )√πœˆπœ‹
2
Γ(
𝑑2
−
(1 + 𝜈 )
1
𝜈+1
2
. When 𝜈 = 1,
1
(1 + 𝑑 2 )−1 =
since Γ(1) = 1 and Γ (2) = √πœ‹. We thus
πœ‹ 1+𝑑 2
1
have that 𝑓𝑇 (𝑑) = πœ‹ 1+𝑑 2 when 𝜈 = 1, which is the density of a Cauchy random variable.
1
𝑑
1
To find the cumulative distribution, we simply compute 𝐹𝑇 (𝑑) = πœ‹ ∫−∞ 1+π‘₯ 2 𝑑π‘₯ =
1
πœ‹
1
πœ‹
1
1
[π‘Žπ‘Ÿπ‘π‘‘π‘Žπ‘›β‘(π‘₯)]𝑑−∞ = (π‘Žπ‘Ÿπ‘π‘‘π‘Žπ‘›(𝑑) − (− )) = π‘Žπ‘Ÿπ‘π‘‘π‘Žπ‘›(𝑑) + .
πœ‹
2
πœ‹
2
b) The 100 ⋅ 𝛾 π‘‘β„Ž percentile is the value of 𝑑 such that 𝐹𝑇 (𝑑) = 𝛾. From the work above,
we have
1
πœ‹
1
1
1
π‘Žπ‘Ÿπ‘π‘‘π‘Žπ‘›(𝑑) + 2 = 𝛾 → π‘Žπ‘Ÿπ‘π‘‘π‘Žπ‘›(𝑑) = (𝛾 − 2) πœ‹ → 𝑑 = π‘‘π‘Žπ‘› [(𝛾 − 2) πœ‹]. This
1
proves that the 100 ⋅ 𝛾 π‘‘β„Ž percentile is given by 𝑑𝛾 (1) = π‘‘π‘Žπ‘› [(𝛾 − 2) πœ‹].
Chapter #8 – Statistics and Sampling Distributions
Question #22: Compute 𝐸(𝑋 𝑝 ) for 𝑝 > 0 if we have that 𝑋~𝐡𝐸𝑇𝐴(π‘Ž, 𝑏).
ο‚·
Γ(π‘Ž+𝑏)
Since 𝑋~𝐡𝐸𝑇𝐴(π‘Ž, 𝑏), its PDF is 𝑓𝑋 (π‘₯) = Γ(π‘Ž)Γ(𝑏) π‘₯ π‘Ž−1 (1 − π‘₯)𝑏−1 whenever 0 < π‘₯ < 1
and π‘Ž > 0, 𝑏 > 0. Then using the definition of expected value, we can compute
∞
1
Γ(π‘Ž+𝑏)
Γ(π‘Ž+𝑏) Γ(π‘Ž+𝑝)Γ(𝑏)
𝐸(𝑋 𝑝 ) = ∫−∞ π‘₯ 𝑝 𝑓𝑋 (π‘₯) 𝑑π‘₯ = Γ(π‘Ž)Γ(𝑏) ∫0 π‘₯ 𝑝 π‘₯ π‘Ž−1 (1 − π‘₯)𝑏−1 𝑑π‘₯ = Γ(π‘Ž)Γ(𝑏)
Γ(π‘Ž+𝑏)Γ(π‘Ž+𝑝)
Γ(π‘Ž)Γ(π‘Ž+𝑏+𝑝)
since we have
integral to conclude that
Γ(π‘Ž+𝑏) 1 π‘Ž−1
∫ π‘₯ (1
Γ(π‘Ž)Γ(𝑏) 0
Γ(π‘Ž+𝑏+𝑝)
=
− π‘₯)𝑏−1 𝑑π‘₯ = 1 so we can solve for the
1
Γ(π‘Ž)Γ(𝑏)
= ∫0 π‘₯ π‘Ž−1 (1 − π‘₯)𝑏−1 𝑑π‘₯. In this case, we are solving
Γ(π‘Ž+𝑏)
1
1
that ∫0 π‘₯ 𝑝 π‘₯ π‘Ž−1 (1 − π‘₯)𝑏−1 𝑑π‘₯ = ∫0 π‘₯ 𝑝+π‘Ž−1 (1 − π‘₯)𝑏−1 𝑑π‘₯ =
Γ(π‘Ž+𝑝)Γ(𝑏)
Γ(π‘Ž+𝑏+𝑝)
. Therefore, all of
the moments of the beta distribution for some fixed 𝑝 > 0 can be written in terms of
the gamma function, which can be evaluated numerically.
Question #24: Suppose that π‘Œπœˆ ~πœ’ 2 (𝜈). Use moment generating functions to find the limiting
distribution of the transformed random variable
ο‚·
π‘Œπœˆ −𝜈
√2𝜈
as 𝜈 → ∞.
This result follows directly from the Central Limit Theorem. If we let π‘Œπœˆ = ∑𝑛𝑖=1 𝑋𝑖
where 𝑋𝑖 ~χ2 (1) for 𝑖 = 1, … , 𝑛, then π‘Œπœˆ ~πœ’ 2 (𝑛) so that 𝐸(π‘Œπœˆ ) = 𝑛, π‘‰π‘Žπ‘Ÿ(π‘Œπœˆ ) = 2𝑛 and
𝑠𝑑(π‘Œπœˆ ) = √2𝑛. Therefore,
π‘Œπœˆ −𝐸(π‘Œπœˆ )
𝑠𝑑(π‘Œπœˆ )
π‘Œπœˆ −𝑛
=
√2𝑛
→ 𝑍~𝑁(0,1) as 𝑛 → ∞. We will now prove
this result using moment generating functions. By the definition of MGFs, we have
𝑀[π‘Œπœˆ−𝜈 ] (𝑑) = 𝐸 [𝑒
π‘Œ −𝜈
𝑑( 𝜈 )
√2𝜈
] = 𝐸 [𝑒
𝜈
2
−𝑑√
𝑒
π‘Œ
𝑑 𝜈
√2𝜈
]=𝑒
−𝑑
√2
√𝜈
𝐸 [𝑒
π‘Œ
𝑑 𝜈
√2𝜈
]=𝑒
−𝑑
√2𝜈
𝑒
√2
−𝑑
√𝜈
(1 −
2𝑑
√2𝜈
−
)
𝜈
2
=𝑒
√2
−𝑑
√𝜈
2
−
√2
√𝜈
𝑑
π‘€π‘Œπœˆ (
√2𝜈
)=
𝜈
2
(1 − 𝑑√𝜈) . In order to evaluate lim 𝑀[π‘Œπœˆ−𝜈 ] (𝑑), we first
𝜈→∞
√2𝜈
take logarithms and then exponentiate the result. This implies that ln [M[Yν−ν ] (t)] =
√2ν
ln [𝑒
−𝑑
√2
√𝜈
(1 − 𝑑
√2
)
√𝜈
𝜈
−
2
] = ln [e
−𝑑
√2
√𝜈
] + ln [(1 − 𝑑
√2
√𝜈
𝜈
−
2
) ] = −𝑑
√2
√𝜈
ν
− 2 ln (1 − 𝑑
√2
√𝜈
). From
here, we use the Taylor series ln(1 − 𝑧) = −𝑧 −
𝑧2
2
−
𝑧3
3
− β‹― for 𝑧 = 𝑑
the limit, which then gives lim ln [M[Yν−ν ] (t)] = lim [−𝑑
𝜈→∞
√2
lim [−𝑑 𝜈
√
𝜈→∞
𝑑2
lim [ 2 +
𝜈→∞
−
ν
√2
(−𝑑 𝜈
2
√
𝑑 3 √2
3√𝜈
+β‹―] =
lim 𝑀[π‘Œπœˆ −𝜈 ] (𝑑) = 𝑒
𝜈→∞
−
𝑑2
𝑑2
2
𝜈
−
3
3𝜈 2
− β‹― )] = lim [−𝑑
𝜈→∞
√2
√𝜈
+𝑑
√2
√𝜈
+
√2
√𝜈
𝑑2
2
to evaluate
ν
− 2 ln (1 − 𝑑
+
𝑑 3 √2
3√𝜈
√2
√𝜈
)] ≈
+β‹―] =
+ 0 + β‹―. This result therefore implies that the limit
lim ln[M Yν −ν (t)]
[
]
𝜈→∞
√2ν
3
𝑑 3 22
𝜈→∞
√2
√𝜈
√2ν
𝑑2
= 𝑒 2 , which is the moment generating function of
√2𝜈
a random variable that follows a standard normal distribution. This proves that the
random variable
π‘Œπœˆ −𝑛
√2𝑛
→ 𝑍~𝑁(0,1) as 𝑛 → ∞, just as is guaranteed by the CLT.
Chapter #9 – Point Estimation
Question #1: Assume that 𝑋1 , … , 𝑋𝑛 are independent and identically distributed with
common density 𝑓(π‘₯; πœƒ), where πœƒ > 0 is an unknown parameter. Find the method of
moments estimator (MME) of πœƒ if the density function is a) 𝑓(π‘₯; πœƒ) = πœƒπ‘₯ πœƒ−1 for 0 < π‘₯ < 1,
b) 𝑓(π‘₯; πœƒ) = (πœƒ + 1)π‘₯ −πœƒ−2 whenever π‘₯ > 1, and c) 𝑓(π‘₯; πœƒ) = πœƒ 2 π‘₯𝑒 −πœƒπ‘₯ whenever π‘₯ > 0.
1
a) We begin by computing the first population moment, so 𝐸(𝑋) = ∫0 π‘₯𝑓(π‘₯; πœƒ) 𝑑π‘₯ =
1
1
1
πœƒ
πœƒ
πœƒ
∫0 π‘₯(πœƒπ‘₯ πœƒ−1 ) 𝑑π‘₯ = πœƒ ∫0 π‘₯ πœƒ 𝑑π‘₯ = πœƒ+1 [π‘₯ πœƒ+1 ]0 = πœƒ+1 (1 − 0) = πœƒ+1. We therefore have
πœƒ
𝐸(𝑋) = πœƒ+1. Next, we equate the first population moment with the first sample
πœƒ
1
πœƒ
moment, which gives πœ‡1′ = 𝑀1′ → πœƒ+1 = 𝑛 ∑𝑛𝑖=1 𝑋𝑖 → πœƒ+1 = 𝑋̅. Finally, we replace πœƒ
Μ‚
πœƒ
𝑋̅
by πœƒΜ‚ and solve the equation πœƒΜ‚+1 = 𝑋̅ for πœƒΜ‚, which implies that πœƒΜ‚π‘€π‘€πΈ = 1−𝑋̅.
∞
∞
b) Just as above, we first compute 𝐸(𝑋) = ∫1 π‘₯𝑓(π‘₯; πœƒ) 𝑑π‘₯ = ∫1 π‘₯[(πœƒ + 1)π‘₯ −πœƒ−2 ] 𝑑π‘₯ =
∞
(πœƒ + 1) ∫1 π‘₯ −πœƒ−1 𝑑π‘₯ =
πœƒ+1
−πœƒ
which means that πœ‡1′ = 𝑀1′ →
∞
∞
[π‘₯ −πœƒ ]1 = −
πœƒ+1
πœƒ
πœƒ+1
πœƒ
1
[0 − 1] =
= 𝑛 ∑𝑛𝑖=1 𝑋𝑖 →
πœƒ+1
πœƒ
πœƒ+1
πœƒ
. Thus, we have 𝐸(𝑋) =
πœƒ+1
πœƒ
1
= 𝑋̅ and πœƒΜ‚π‘€π‘€πΈ = 𝑋̅−1.
∞
∞
2
c) We have 𝐸(𝑋) = ∫0 π‘₯𝑓(π‘₯; πœƒ) 𝑑π‘₯ = ∫0 π‘₯[πœƒ 2 π‘₯𝑒 −πœƒπ‘₯ ] 𝑑π‘₯ = πœƒ 2 ∫0 π‘₯ 2 𝑒 −πœƒπ‘₯ 𝑑π‘₯ = β‹― = πœƒ
after doing integration by parts. We can also find this directly by noting that the
1
density 𝑓(π‘₯, πœƒ) = πœƒ 2 π‘₯𝑒 −πœƒπ‘₯ suggests that 𝑋~𝐺𝐴𝑀𝑀𝐴 (πœƒ , 2). This then implies that
1
2
2
1
2
𝐸(𝑋) = πœ…πœƒ = πœƒ 2 = πœƒ. We therefore set πœ‡1′ = 𝑀1′ such that πœƒ = 𝑛 ∑𝑛𝑖=1 𝑋𝑖 or πœƒ = 𝑋̅, and
2
then solve for the method of moments estimator, which is given by πœƒΜ‚π‘€π‘€πΈ = Μ… .
𝑋
Question #2: Assume that 𝑋1 , … , 𝑋𝑛 are independent and identically distributed. Find the
method of moments estimator (MME) of the unknown parameters if the random sample
1
comes from a) 𝑋~𝑁𝐡(3, 𝑝), b) 𝑋~𝐺𝐴𝑀𝑀𝐴(2, πœ…), c) 𝑋~π‘ŠπΈπΌ (πœƒ, 2), and d) 𝑋~𝑃𝐴𝑅(πœƒ, πœ…).
π‘Ÿ
3
a) Since 𝑋~𝑁𝐡(3, 𝑝), we know that 𝐸(𝑋) = 𝑝 = 𝑝. Equating this with the first sample
3
3
moment gives πœ‡1′ = 𝑀1′ → 𝑝 = 𝑋̅, so the estimator is 𝑝̂𝑀𝑀𝐸 = 𝑋̅.
b) Since 𝑋~𝐺𝐴𝑀𝑀𝐴(2, πœ…), we know that 𝐸(𝑋) = πœ…πœƒ = 2πœ…. Equating this with the first
𝑋̅
sample moment gives πœ‡1′ = 𝑀1′ → 2πœ… = 𝑋̅, so the estimator is πœ…Μ‚ 𝑀𝑀𝐸 = 2 .
1
1
1
c) Since 𝑋~π‘ŠπΈπΌ (πœƒ, 2), we know that 𝐸(𝑋) = πœƒΓ (1 + 𝛽) = πœƒΓ (1 + 1/2) = πœƒΓ(3) =
𝑋̅
πœƒ(3 − 1)! = 2πœƒ. Thus, we have πœ‡1′ = 𝑀1′ → 2πœƒ = 𝑋̅, so the estimator is πœƒΜ‚π‘€π‘€πΈ = 2 .
πœƒ2 πœ…
πœƒ
πœƒ2
d) Since 𝑋~𝑃𝐴𝑅(πœƒ, πœ…), we have πœ‡1 = πœ…−1 and πœ‡2 = 𝜎 2 + πœ‡12 = (πœ…−2)(πœ…−1)2 + (πœ…−1)2. This
πœƒ2 πœ…
πœƒ
πœƒ2
1
means that πœ‡1 = πœ…−1 = 𝑀1′ = 𝑋̅ and πœ‡2 = (πœ…−2)(πœ…−1)2 + (πœ…−1)2 = 𝑀2′ = 𝑛 ∑𝑛𝑖=1 𝑋𝑖2 . We
must solve for the unknown parameters πœƒ and πœ… in terms of the two sample moments
𝑋̅ and
1
𝑛
∑𝑛𝑖=1 𝑋𝑖2 . From the first equation, we can solve to find πœƒ = (πœ… − 1)𝑋̅ and
substitute into the second equation to find
𝑋̅ 2 πœ…
(πœ…−2)
𝑋̅ 2 (πœ…−1)2 πœ…
(πœ…−2)(πœ…−1)2
+
𝑋̅ 2 (πœ…−1)2
(πœ…−1)2
𝑛
1
= 𝑛 ∑𝑛𝑖=1 𝑋𝑖2 →
2
∑
𝑋𝑖
1
πœ…
1
2πœ…−2
+ 𝑋̅ 2 = 𝑛 ∑𝑛𝑖=1 𝑋𝑖2 → 𝑋̅ 2 (πœ…−2 + 1) = 𝑛 ∑𝑛𝑖=1 𝑋𝑖2 → πœ…−2 = 𝑖=1
. But this means
2
Μ…
𝑛𝑋
𝑛𝑋̅ 2 (2πœ… − 2) = (πœ… − 2) ∑𝑛𝑖=1 𝑋𝑖2 → 2𝑛𝑋̅ 2 πœ… − 2𝑛𝑋̅ 2 = πœ… ∑𝑛𝑖=1 𝑋𝑖2 − 2 ∑𝑛𝑖=1 𝑋𝑖2 , so that
2𝑛𝑋̅ 2 πœ… − πœ… ∑𝑛𝑖=1 𝑋𝑖2 = 2𝑛𝑋̅ 2 − 2 ∑𝑛𝑖=1 𝑋𝑖2 → πœ…(2𝑛𝑋̅ 2 − ∑𝑛𝑖=1 𝑋𝑖2 ) = 2𝑛𝑋̅ 2 − 2 ∑𝑛𝑖=1 𝑋𝑖2 .
Finally, we divide through to find πœ… =
2
2𝑛𝑋̅ 2 −2 ∑𝑛
𝑖=1 𝑋𝑖
2 .
2𝑛𝑋̅ 2 −∑𝑛
𝑖=1 𝑋𝑖
Plugging in to the other equation
2𝑛𝑋̅ 2 −2 ∑𝑛 𝑋 2
implies that πœƒ = (πœ… − 1)𝑋̅ = ( 2𝑛𝑋̅ 2 −∑𝑛𝑖=1𝑋 2𝑖 − 1) 𝑋̅, so that the two method of
𝑖=1 𝑖
moments estimators are πœ…Μ‚ 𝑀𝑀𝐸 =
2
2𝑛𝑋̅ 2 −2 ∑𝑛
𝑖=1 𝑋𝑖
𝑛
2
2
2𝑛𝑋̅ −∑𝑖=1 𝑋𝑖
2𝑛𝑋̅ 2 −2 ∑𝑛
𝑋2
and πœƒΜ‚π‘€π‘€πΈ = ( 2𝑛𝑋̅ 2 −∑𝑛𝑖=1𝑋 2𝑖 − 1) 𝑋̅.
𝑖=1 𝑖
Question #3: Assume that 𝑋1 , … , 𝑋𝑛 are independent and identically distributed with
common density 𝑓(π‘₯; πœƒ), where πœƒ > 0 is an unknown parameter. Find the maximum
likelihood estimator (MLE) for πœƒ when the PDF is a) 𝑓(π‘₯; πœƒ) = πœƒπ‘₯ πœƒ−1 whenever 0 < π‘₯ < 1,
b) 𝑓(π‘₯; πœƒ) = (πœƒ + 1)π‘₯ −πœƒ−2 whenever π‘₯ > 1, and c) 𝑓(π‘₯, πœƒ) = πœƒ 2 π‘₯𝑒 −πœƒπ‘₯ whenever π‘₯ > 0.
a) We first find the likelihood function based on the joint density of 𝑋1 , … , 𝑋𝑛 , which is
𝐿(πœƒ) = 𝑓(π‘₯1 ; πœƒ) … 𝑓(π‘₯𝑛 ; πœƒ) = ∏𝑛𝑖=1 𝑓(π‘₯𝑖 ; πœƒ) = ∏𝑛𝑖=1 πœƒπ‘₯π‘–πœƒ−1 = πœƒ 𝑛 (π‘₯1 … π‘₯𝑛 )πœƒ−1 . Next, we
construct the log likelihood function, since it is easier to differentiate and achieves a
maximum at the same point as the likelihood function. This gives ln[𝐿(πœƒ)] =
ln[πœƒ 𝑛 (π‘₯1 … π‘₯𝑛 )πœƒ−1 ] = 𝑛 ln(πœƒ) + (πœƒ − 1)[ln(π‘₯1 ) + β‹― + ln(π‘₯𝑛 )], which we differentiate
so
πœ•
πœ•
𝑛
ln[𝐿(πœƒ)] = πœ•πœƒ [𝑛 ln(πœƒ) + (πœƒ − 1) ∑𝑛𝑖=1 ln(π‘₯𝑖 )] = πœƒ + ∑𝑛𝑖=1 ln(π‘₯𝑖 ). We then solve
πœ•πœƒ
𝑛
for the value of πœƒ which makes the derivative equal zero, so πœƒ + ∑𝑛𝑖=1 ln(π‘₯𝑖 ) = 0 → πœƒ =
− ∑𝑛
𝑛
. Since it is clear that the second derivative of ln[𝐿(πœƒ)] is negative, we have
𝑖=1 ln(π‘₯𝑖 )
found that the maximum likelihood estimator is πœƒΜ‚π‘€πΏπΈ = − ∑𝑛
𝑛
. (Note that we
𝑖=1 ln(𝑋𝑖 )
must capitalize the 𝑋𝑖 from π‘₯𝑖 when presenting the estimator.)
b) We have 𝐿(πœƒ) = ∏𝑛𝑖=1 𝑓(π‘₯𝑖 ; πœƒ) = ∏𝑛𝑖=1(πœƒ + 1)π‘₯𝑖−πœƒ−2 = (πœƒ + 1)𝑛 (π‘₯1 … π‘₯𝑛 )−πœƒ−2 so that
ln[𝐿(πœƒ)] = ln[(πœƒ + 1)𝑛 (π‘₯1 … π‘₯𝑛 )−πœƒ−2 ] = 𝑛 ln(πœƒ + 1) − (πœƒ + 2) ∑𝑛𝑖=1 ln(π‘₯𝑖 ). Then we
find
πœ•
πœ•πœƒ
πœ•
𝑛
ln[𝐿(πœƒ)] = πœ•πœƒ [𝑛 ln(πœƒ + 1) − (πœƒ + 2) ∑𝑛𝑖=1 ln(π‘₯𝑖 )] = πœƒ+1 − ∑𝑛𝑖=1 ln(π‘₯𝑖 ). Finally,
𝑛
we must solve πœƒ+1 − ∑𝑛𝑖=1 ln(π‘₯𝑖 ) = 0 → πœƒ = ∑𝑛
𝑛
𝑖=1 ln(π‘₯𝑖 )
− 1. Since the second derivative
of ln[𝐿(πœƒ)] will be negative, we have found that πœƒΜ‚π‘€πΏπΈ = ∑𝑛
𝑛
𝑖=1 ln(𝑋𝑖 )
− 1.
c) We have 𝐿(πœƒ) = ∏𝑛𝑖=1 𝑓(π‘₯𝑖 ; πœƒ) = ∏𝑛𝑖=1 πœƒ 2 π‘₯𝑖 𝑒 −πœƒπ‘₯𝑖 = πœƒ 2𝑛 (π‘₯1 … π‘₯𝑛 )𝑒 −πœƒ(π‘₯1 +β‹―+π‘₯𝑛) so that
𝑛
ln[𝐿(πœƒ)] = ln[πœƒ 2𝑛 (π‘₯1 … π‘₯𝑛 )𝑒 −πœƒ(∑𝑖=1 π‘₯𝑖 ) ] = 2𝑛 ln(πœƒ) + ∑𝑛𝑖=1 ln(π‘₯𝑖 ) − πœƒ ∑𝑛𝑖=1 π‘₯𝑖 . Then we
have
πœ•
πœ•πœƒ
πœ•
ln[𝐿(πœƒ)] = πœ•πœƒ [2𝑛 ln(πœƒ) + ∑𝑛𝑖=1 ln(π‘₯𝑖 ) − πœƒ ∑𝑛𝑖=1 π‘₯𝑖 ] =
must solve
2𝑛
πœƒ
2𝑛
− ∑𝑛𝑖=1 π‘₯𝑖 = 0 → πœƒ = ∑𝑛
𝑖=1 π‘₯𝑖
2𝑛
πœƒ
− ∑𝑛𝑖=1 π‘₯𝑖 . Finally, we
2
2
= π‘₯Μ… , which implies that πœƒΜ‚π‘€πΏπΈ = 𝑋̅.
Question #4: Assume that 𝑋1 , … , 𝑋𝑛 are independent and identically distributed. Find the
maximum likelihood estimator (MLE) of the parameter if the distribution is a) 𝑋𝑖 ~𝐡𝐼𝑁(1, 𝑝),
1
b) 𝑋𝑖 ~𝐺𝐸𝑂(𝑝) , c) 𝑋𝑖 ~𝑁𝐡(3, 𝑝), d) 𝑋𝑖 ~𝐺𝐴𝑀𝑀𝐴(πœƒ, 2), e) 𝑋𝑖 ~π‘ŠπΈπΌ (πœƒ, 2), and f) 𝑋𝑖 ~𝑃𝐴𝑅(1, πœ…).
a) Since the density of 𝑋~𝐡𝐼𝑁(1, 𝑝) is 𝑓(π‘₯; 𝑝) = (π‘₯1)𝑝 π‘₯ (1 − 𝑝)1−π‘₯ = 𝑝 π‘₯ (1 − 𝑝)1−π‘₯ , we
𝑛
𝑛
have 𝐿(𝑝) = ∏𝑛𝑖=1 𝑓(π‘₯𝑖 ; 𝑝) = ∏𝑛𝑖=1 𝑝 π‘₯𝑖 (1 − 𝑝)1−π‘₯𝑖 = 𝑝∑𝑖=1 π‘₯𝑖 (1 − 𝑝)𝑛−∑𝑖=1 π‘₯𝑖 and then
𝑛
𝑛
ln[𝐿(𝑝)] = ln[𝑝∑𝑖=1 π‘₯𝑖 (1 − 𝑝)𝑛−∑𝑖=1 π‘₯𝑖 ] = (∑𝑛𝑖=1 π‘₯𝑖 ) ln(𝑝) + (𝑛 − ∑𝑛𝑖=1 π‘₯𝑖 )ln(1 − 𝑝).
Differentiating
∑𝑛
𝑖=1 π‘₯𝑖
𝑝
−
𝑛−∑𝑛
𝑖=1 π‘₯𝑖
1−𝑝
gives
=0→
πœ•
πœ•π‘
ln[𝐿(𝑝)] =
∑𝑛
𝑖=1 π‘₯𝑖
𝑝
=
πœ•
πœ•π‘
𝑛−∑𝑛
𝑖=1 π‘₯𝑖
1−𝑝
[(∑𝑛𝑖=1 π‘₯𝑖 ) ln(𝑝) + (𝑛 − ∑𝑛𝑖=1 π‘₯𝑖 )ln(1 − 𝑝)] =
→ (1 − 𝑝) ∑𝑛𝑖=1 π‘₯𝑖 = 𝑝(𝑛 − ∑𝑛𝑖=1 π‘₯𝑖 ) →
∑𝑛𝑖=1 π‘₯𝑖 − 𝑝 ∑𝑛𝑖=1 π‘₯𝑖 = 𝑝𝑛 − 𝑝 ∑𝑛𝑖=1 π‘₯𝑖 → ∑𝑛𝑖=1 π‘₯𝑖 = 𝑝𝑛 → 𝑝 =
∑𝑛
𝑖=1 π‘₯𝑖
𝑛
= π‘₯Μ… .
Since
the
second derivative will be negative, we have found that 𝑝̂ 𝑀𝐿𝐸 = 𝑋̅.
b) Since 𝑓(π‘₯; 𝑝) = 𝑝(1 − 𝑝)π‘₯−1 , we have 𝐿(𝑝) = ∏𝑛𝑖=1 𝑓(π‘₯𝑖 ; 𝑝) = ∏𝑛𝑖=1 𝑝(1 − 𝑝)π‘₯𝑖 −1 =
𝑛
𝑝𝑛 (1 − 𝑝)[∑𝑖=1 π‘₯𝑖 ]−𝑛 and then the log likelihood function becomes ln[𝐿(𝑝)] =
𝑛
ln[𝑝𝑛 (1 − 𝑝)[∑𝑖=1 π‘₯𝑖 ]−𝑛 ] = 𝑛 ln(𝑝) + {[∑𝑛𝑖=1 π‘₯𝑖 ] − 𝑛}ln(1 − 𝑝).
πœ•
πœ•
𝑛
ln[𝐿(𝑝)] = πœ•π‘ [𝑛 ln(𝑝) + {[∑𝑛𝑖=1 π‘₯𝑖 ] − 𝑛} ln(1 − 𝑝)] = 𝑝 −
πœ•π‘
with zero implies
𝑛
−
𝑝
[∑𝑛
𝑖=1 π‘₯𝑖 ]−𝑛
1−𝑝
𝑛
=0→𝑝=
[∑𝑛
𝑖=1 π‘₯𝑖 ]−𝑛
1−𝑝
𝑛
𝑛 − 𝑛𝑝 = 𝑝 ∑𝑛𝑖=1 π‘₯𝑖 − 𝑛𝑝 → 𝑛 = 𝑝 ∑𝑛𝑖=1 π‘₯𝑖 → 𝑝 = ∑𝑛
𝑖=1 π‘₯𝑖
Differentiating
[∑𝑛
𝑖=1 π‘₯𝑖 ]−𝑛
1−𝑝
gives
. Equating this
→ (1 − 𝑝)𝑛 = 𝑝[∑𝑛𝑖=1 π‘₯𝑖 − 𝑛] →
=
1
1 𝑛
∑
π‘₯
𝑛 𝑖=1 𝑖
1
= π‘₯Μ… . Since the second
1
derivative will be negative, we have found that 𝑝̂ 𝑀𝐿𝐸 = 𝑋̅.
(π‘₯−1)!
c) Since 𝑋~𝑁𝐡(3, 𝑝), we have 𝑓(π‘₯; 𝑝) = (π‘₯−1
)𝑝3 (1 − 𝑝)π‘₯−3 = 2(π‘₯−3)! 𝑝3 (1 − 𝑝)π‘₯−3 =
3−1
(π‘₯−1)(π‘₯−2)
2
1
𝑝3 (1 − 𝑝)π‘₯−3 = 2 (π‘₯ 2 − 3π‘₯ + 2)𝑝3 (1 − 𝑝)π‘₯−3.
This
implies
that
the
1
likelihood function 𝐿(𝑝) = ∏𝑛𝑖=1 𝑓(π‘₯𝑖 ; 𝑝) = ∏𝑛𝑖=1 [2 (π‘₯𝑖2 − 3π‘₯𝑖 + 2)𝑝3 (1 − 𝑝)π‘₯𝑖 −3 ] =
𝑛
2−𝑛 (π‘₯𝑖2 − 3π‘₯𝑖 + 2)𝑛 𝑝3𝑛 (1 − 𝑝)[∑𝑖=1 π‘₯𝑖 ]−3𝑛 , so the log likelihood function ln[𝐿(𝑝)] =
𝑛
ln[2−𝑛 (π‘₯𝑖2 − 3π‘₯𝑖 + 2)𝑛 𝑝3𝑛 (1 − 𝑝)[∑𝑖=1 π‘₯𝑖 ]−3𝑛 ] = −𝑛 ln(2) + 𝑛 ln(π‘₯𝑖2 − 3π‘₯𝑖 + 2) +
3𝑛 ln(𝑝) + {[∑𝑛𝑖=1 π‘₯𝑖 ] − 3𝑛} ln(1 − 𝑝). Differentiating this then gives
πœ•
πœ•π‘
πœ•
ln[𝐿(𝑝)] =
πœ•π‘
[−𝑛 ln(2) + 𝑛 ln(π‘₯𝑖2 − 3π‘₯𝑖 + 2) + 3𝑛 ln(𝑝) + {[∑𝑛𝑖=1 π‘₯𝑖 ] − 3𝑛} ln(1 − 𝑝)] =
[∑𝑛
𝑖=1 π‘₯𝑖 ]−3𝑛
1−𝑝
3
3𝑛
𝑝
−
3
= 0 → β‹― → 𝑝 = 𝑋̅. Therefore, we have that 𝑝̂ 𝑀𝐿𝐸 = 𝑋̅.
π‘₯
1
π‘₯
1
d) Since 𝑋~𝐺𝐴𝑀𝑀𝐴(πœƒ, 2), we have 𝑓(π‘₯; πœƒ) = πœƒ2 Γ(2) π‘₯ 2−1 𝑒 −πœƒ = πœƒ2 π‘₯𝑒 −πœƒ . This means
π‘₯𝑖
1
1
1
𝐿(πœƒ) = ∏𝑛𝑖=1 𝑓(π‘₯𝑖 ; πœƒ) = ∏𝑛𝑖=1 [πœƒ2 π‘₯𝑖 𝑒 − πœƒ ] = πœƒ2𝑛 (π‘₯1 … π‘₯𝑛 )𝑒 −πœƒ
1
1
ln [πœƒ2𝑛 (π‘₯1 … π‘₯𝑛 )𝑒 −πœƒ
gives
πœ•
∑𝑛
𝑖=1 π‘₯𝑖
so that ln[𝐿(πœƒ)] =
1
] = −2𝑛 ln(πœƒ) + ∑𝑛𝑖=1 ln(π‘₯𝑖 ) − πœƒ ∑𝑛𝑖=1 π‘₯𝑖 .
πœ•
πœ•πœƒ
∑𝑛
𝑖=1 π‘₯𝑖
1
ln[𝐿(πœƒ)] = πœ•πœƒ [−2𝑛 ln(πœƒ) + ∑𝑛𝑖=1 ln(π‘₯𝑖 ) − πœƒ ∑𝑛𝑖=1 π‘₯𝑖 ] = −
we solve
1
πœƒ2
∑𝑛𝑖=1 π‘₯𝑖 −
2𝑛
πœƒ
1
= 0 → πœƒ2 ∑𝑛𝑖=1 π‘₯𝑖 =
2𝑛
πœƒ
Differentiating
2𝑛
πœƒ
1
+ πœƒ2 ∑𝑛𝑖=1 π‘₯𝑖 . Then
→ πœƒ ∑𝑛𝑖=1 π‘₯𝑖 = 2π‘›πœƒ 2 → πœƒ =
∑𝑛
𝑖=1 π‘₯𝑖
2𝑛
π‘₯Μ…
= 2.
𝑋̅
Since the second derivative will be zero, we have found that πœƒΜ‚π‘€πΏπΈ = 2 .
1
1
e) Since 𝑋~π‘ŠπΈπΌ (πœƒ, 2), we have 𝑓(π‘₯; πœƒ) = 2√πœƒ π‘₯
𝐿(πœƒ) =
∏𝑛𝑖=1 𝑓(π‘₯𝑖 ; πœƒ)
1
=
1
−1
2
π‘₯
− −√ 𝑖
1
2
∏𝑛𝑖=1 [
π‘₯
𝑒 √πœƒ ]
2√πœƒ 𝑖
=
1
𝑒
π‘₯
πœƒ
−√
𝑛
2𝑛 πœƒ 2
−
𝑛
1
πœ•
−
𝑛
1
√πœƒ
1
𝑛
√πœƒ =
∑𝑛
𝑖=1 √π‘₯𝑖
𝑛
→πœƒ=[
∑𝑛
𝑖=1 √π‘₯𝑖
𝑛
2
1
2
(π‘₯1
1
2
π‘₯
πœƒ
−√
. Thus, we have
1
1
1
−
− ∑𝑛 π‘₯ 2
… π‘₯𝑛 2 ) 𝑒 √πœƒ 𝑖=1 𝑖
1
−
𝑛
2𝑛 πœƒ 2
(π‘₯1
1
2
√πœƒ
∑𝑛
𝑖=1 √π‘₯𝑖
3
2πœƒ2
1
∑𝑛𝑖=1 π‘₯𝑖2 ] = −
=0→
𝑛
+
2πœƒ
∑𝑛
𝑖=1 √π‘₯𝑖
3
2πœƒ2
∑𝑛
𝑖=1 √π‘₯𝑖
3
2πœƒ2
1
1
1
1
so that the
1
−
− ∑𝑛 π‘₯ 2
… π‘₯𝑛 2 ) 𝑒 √πœƒ 𝑖=1 𝑖 ]
∑𝑛𝑖=1 π‘₯𝑖2 . Differentiating this gives
[−𝑛 ln(2) − 2 ln(πœƒ) + ∑𝑛𝑖=1 π‘₯𝑖 2 −
πœ•πœƒ
zero and solving implies − 2πœƒ +
−
= 2√πœƒ π‘₯ 𝑒
−
log of the likelihood function is ln[𝐿(πœƒ)] = ln [
−𝑛 ln(2) − 2 ln(πœƒ) + ∑𝑛𝑖=1 π‘₯𝑖 2 −
1
πœ•
πœ•πœƒ
=
ln[𝐿(πœƒ)] =
. Setting this equal to
3
𝑛
= 2πœƒ → 2πœƒ ∑𝑛𝑖=1 √π‘₯𝑖 = 2π‘›πœƒ 2 →
𝑛
2
∑
√𝑋
] . Therefore, we have found πœƒΜ‚π‘€πΏπΈ = [ 𝑖=1𝑛 𝑖 ] .
πœ…
f) Since 𝑋~𝑃𝐴𝑅(1, πœ…), we have 𝑓(π‘₯; πœ…) = (1+π‘₯)πœ…+1 so the likelihood function is 𝐿(πœ…) =
∏𝑛𝑖=1 𝑓(π‘₯𝑖 ; πœ…) = ∏𝑛𝑖=1[πœ…(1 + π‘₯𝑖 )−πœ…−1 ] = πœ… 𝑛 ∏𝑛𝑖=1(1 + π‘₯𝑖 )−πœ…−1. Then we have that
ln[𝐿(πœ…)] = ln[πœ… 𝑛 ∏𝑛𝑖=1(1 + π‘₯𝑖 )−πœ…−1 ] = 𝑛 ln(πœ…) − (πœ… + 1) ∑𝑛𝑖=1 ln(1 + π‘₯𝑖 ).
compute the derivative so that
𝑛
πœ…
𝑛
πœ…
πœ•
πœ•πœ…
Next,
we
πœ•
ln[𝐿(πœ…)] = πœ•πœ… [𝑛 ln(πœ…) − (πœ… + 1) ∑𝑛𝑖=1 ln(1 + π‘₯𝑖 )] =
− ∑𝑛𝑖=1 ln(1 + π‘₯𝑖 ). Finally, we set this result equal to zero and solve for πœ… to find that
𝑛
𝑛
− ∑𝑛𝑖=1 ln(1 + π‘₯𝑖 ) = 0 → πœ… = ∑𝑛𝑖=1 ln(1 + π‘₯𝑖 ) → πœ… = ∑𝑛
. Since the second
𝑖=1 ln(1+π‘₯𝑖 )
derivative will be negative, we have found that πœ…Μ‚ 𝑀𝐿𝐸 = ∑𝑛
𝑛
.
𝑖=1 ln(1+𝑋𝑖 )
Chapter #9 – Point Estimation
Question #7: Let 𝑋1 , … , 𝑋𝑛 be a random sample from 𝑋~𝐺𝐸𝑂(𝑝). Find the Maximum
1
Likelihood Estimator (MLE) for a) 𝐸(𝑋) = 𝑝, b) π‘‰π‘Žπ‘Ÿ(𝑋) =
1−𝑝
𝑝2
, and c) 𝑃(𝑋 > π‘˜) = (1 − 𝑝)π‘˜
where π‘˜ ∈ {1,2, … }. Do it both ways for each part to verify the Invariance Property.
a) We begin by computing 𝑝̂𝑀𝐿𝐸 by first calculating the likelihood function 𝐿(𝑝) =
𝑛
∏𝑛𝑖=1 𝑓(π‘₯𝑖 , 𝑝) = ∏𝑛𝑖=1(1 − 𝑝)π‘₯𝑖 −1 𝑝 = 𝑝𝑛 (1 − 𝑝)∑𝑖=1(π‘₯𝑖 −1). Then we can compute
𝑛
ln[𝐿(𝑝)] = ln[𝑝𝑛 (1 − 𝑝)∑𝑖=1(π‘₯𝑖 −1) ] = 𝑛 ln(𝑝) + [∑𝑛𝑖=1(π‘₯𝑖 − 1)] ln(1 − 𝑝)
πœ•
πœ•
and
𝑛
ln[𝐿(𝑝)] = πœ•π‘ [𝑛 ln(𝑝) + [∑𝑛𝑖=1(π‘₯𝑖 − 1)] ln(1 − 𝑝)] = 𝑝 −
πœ•π‘
differentiate
Setting equal to zero and solving for 𝑝 gives
𝑛
−
𝑝
∑𝑛
𝑖=1 π‘₯𝑖 −𝑛
1−𝑝
𝑛
=0→𝑝=
(1 − 𝑝)𝑛 = 𝑝 ∑𝑛𝑖=1 π‘₯𝑖 − 𝑛𝑝 → 𝑛 − 𝑝𝑛 = 𝑝 ∑𝑛𝑖=1 π‘₯𝑖 − 𝑝𝑛 → 𝑛 = 𝑝 ∑𝑛𝑖=1 π‘₯𝑖 .
𝑛
implies that 𝑝 = ∑𝑛
𝑖=1 π‘₯𝑖
then
∑𝑛
𝑖=1(π‘₯𝑖 −1)
1−𝑝
∑𝑛
𝑖=1 π‘₯𝑖 −𝑛
1−𝑝
This
.
→
then
1
= π‘₯Μ… . Since the second derivative will be negative, we have
1
found that 𝑝̂𝑀𝐿𝐸 = 𝑋̅. By the Invariance Property of the Maximum Likelihood
Estimator, we have that 𝜏(𝑝̂ ) =
1
1
𝑝̂𝑀𝐿𝐸
=
b) Since 𝑝̂ 𝑀𝐿𝐸 = 𝑋̅ and 𝜏(𝑝) = π‘‰π‘Žπ‘Ÿ(𝑋) =
1
1/𝑋̅
1−𝑝
𝑝2
1
= 𝑋̅ as the MLE for 𝜏(𝑝) = 𝐸(𝑋) = .
, then 𝜏(𝑝̂ ) =
𝑝
1−𝑝̂𝑀𝐿𝐸
𝑝̂𝑀𝐿𝐸 2
Μ…
1−1/𝑋
= (1/𝑋̅)2 = 𝑋̅(𝑋̅ − 1), by
the Invariance Property of the Maximum Likelihood Estimator.
1
c) Since 𝑝̂ 𝑀𝐿𝐸 = 𝑋̅ and 𝜏(𝑝) = (1 − 𝑝)π‘˜ , then 𝜏(𝑝̂ ) = (1 − 𝑝̂ 𝑀𝐿𝐸 )π‘˜ = (1 − 1/𝑋̅)π‘˜ , by the
Invariance Property of the Maximum Likelihood Estimator.
Question #12: Let 𝑋1 , … , 𝑋𝑛 be a random sample from 𝑋~𝐿𝑂𝐺𝑁(πœ‡, 𝜎 2 ). Find the Maximum
Likelihood Estimator (MLE) for a) the parameters πœ‡ and 𝜎 2 , and b) 𝜏(πœ‡, 𝜎 2 ) = 𝐸(𝑋).
1
1
a) We have that the density function of 𝑋 is 𝑓𝑋 (π‘₯; πœ‡, 𝜎 2 ) = √2πœ‹πœŽ2 π‘₯ 𝑒
−
1
(ln π‘₯−πœ‡)2
2𝜎2
, so that
the likelihood function of the sample is given by 𝐿(πœ‡, 𝜎 2 ) = ∏𝑛𝑖=1 𝑓(π‘₯𝑖 , 𝑝) =
∏𝑛𝑖=1 [
1
1
√2πœ‹πœŽ2 π‘₯𝑖
𝑒
−
1
(ln π‘₯𝑖 −πœ‡)2
2𝜎2
𝑛
] = (2πœ‹πœŽ 2 )− 2 (π‘₯1 … π‘₯𝑛 )−1 𝑒
−
1
∑𝑛 (ln π‘₯𝑖 −πœ‡)2
2𝜎2 𝑖=1
𝑛
likelihood function is ln[𝐿(πœ‡, 𝜎 2 )] = ln [(2πœ‹πœŽ 2 )− 2 (π‘₯1 … π‘₯𝑛 )−1 𝑒
𝑛
−
. Then the log
1
∑𝑛 (ln π‘₯𝑖 −πœ‡)2
2𝜎2 𝑖=1
]=
1
− 2 ln(2πœ‹πœŽ 2 ) − ∑𝑛𝑖=1 ln(π‘₯𝑖 ) − 2𝜎2 ∑𝑛𝑖=1(ln π‘₯𝑖 − πœ‡)2. We differentiate this with respect
to both parameters and set the resulting expressions equal to zero so we can
simultaneously solve for the parameters, so
and
πœ•
πœ•πœŽ2
𝑛
πœ•
πœ•πœ‡
1
ln[𝐿(πœ‡, 𝜎 2 )] = 𝜎2 ∑𝑛𝑖=1(ln π‘₯𝑖 − πœ‡) = 0
1
ln[𝐿(πœ‡, 𝜎 2 )] = − 2𝜎2 + 2𝜎4 ∑𝑛𝑖=1(ln π‘₯𝑖 − πœ‡)2 = 0. The first equation implies
1
∑𝑛 (ln π‘₯𝑖 − πœ‡) = 0 → ∑𝑛𝑖=1(ln π‘₯𝑖 − πœ‡) = 0 → ∑𝑛𝑖=1(ln π‘₯𝑖 ) − π‘›πœ‡ = 0 → πœ‡ =
𝜎2 𝑖=1
and
1
𝜎2
the
second
𝑛
1
∑𝑛
𝑖=1(ln π‘₯𝑖 )
𝑛
1
𝑛
− 2𝜎2 + 2𝜎4 ∑𝑛𝑖=1(ln π‘₯𝑖 − πœ‡)2 = 0 → 2𝜎4 ∑𝑛𝑖=1(ln π‘₯𝑖 − πœ‡)2 = 2𝜎2 →
1
∑𝑛𝑖=1(ln π‘₯𝑖 − πœ‡)2 = 𝑛 → 𝜎 2 = ∑𝑛𝑖=1(ln π‘₯𝑖 − πœ‡)2. Thus, we have that the maximum
𝑛
1
1
2
likelihood estimators are πœ‡Μ‚ 𝑀𝐿𝐸 = 𝑛 ∑𝑛𝑖=1(ln π‘₯𝑖 ) and πœŽΜ‚π‘€πΏπΈ
= 𝑛 ∑𝑛𝑖=1(ln π‘₯𝑖 − πœ‡Μ‚ 𝑀𝐿𝐸 )2.
b) We know that 𝑋~𝐿𝑂𝐺𝑁(πœ‡, 𝜎 2 ) if and only if π‘Œ = ln(𝑋) ~𝑁(πœ‡, 𝜎 2 ). But π‘Œ = ln⁑(𝑋) if and
only if 𝑋 = 𝑒 π‘Œ implies that 𝐸(𝑋) = 𝐸(𝑒 π‘Œ ) = π‘€π‘Œ (1) = 𝑒 πœ‡(1)+
𝜎2 12
2
𝜎2
= 𝑒 πœ‡+ 2 . By the
Invariance Property of the Maximum Likelihood Estimator, we can conclude that
𝜎2
1 2
2 )
𝜏(πœ‡Μ‚ , πœŽΜ‚ 2 )𝑀𝐿𝐸 = 𝜏(πœ‡Μ‚ 𝑀𝐿𝐸 , πœŽΜ‚π‘€πΏπΈ
= 𝑒 πœ‡Μ‚π‘€πΏπΈ +2πœŽΜ‚π‘€πΏπΈ is the MLE for 𝜏(πœ‡, 𝜎 2 ) = 𝐸(𝑋) = 𝑒 πœ‡+ 2 .
Question #17: Let 𝑋1 , … , 𝑋𝑛 be a random sample from 𝑋~π‘ˆπ‘πΌπΉ(πœƒ − 1, πœƒ + 1). a) Show that
the sample mean 𝑋̅ is an unbiased estimator for πœƒ; b) show that the midrange 𝑀 =
𝑋(1) +𝑋(𝑛)
2
is also an unbiased estimator for the parameter πœƒ; c) which one has a smaller variance?
a) To show that 𝑋̅ is an unbiased estimator for πœƒ, we must verify that 𝐸(𝑋̅) = πœƒ. But we
(πœƒ+1)+(πœƒ−1)
1
1
1
1
see that 𝐸(𝑋̅) = 𝐸 (𝑛 ∑𝑛𝑖=1 𝑋𝑖 ) = 𝑛 𝐸(∑𝑛𝑖=1 𝑋𝑖 ) = 𝑛 ∑𝑛𝑖=1 𝐸(𝑋𝑖 ) = 𝑛 ∑𝑛𝑖=1 [
]=
2
1
1
∑𝑛 πœƒ = π‘›πœƒ = πœƒ, so it is clear that the sample mean is an unbiased estimator for πœƒ.
𝑛 𝑖=1
𝑛
𝑋(1) +𝑋(𝑛)
b) We have that 𝐸(𝑀) = 𝐸 (
2
1
1
) = 2 𝐸(𝑋(1) + 𝑋(𝑛) ) = 2 [𝐸(𝑋(1) ) + 𝐸(𝑋(𝑛) )]. We
must therefore compute the mean of the smallest and largest order statistics, which
we can do by first finding their density functions. We first note that since
1
1
𝑋~π‘ˆπ‘πΌπΉ(πœƒ − 1, πœƒ + 1), then 𝑓𝑋 (𝑑) = (πœƒ+1)−(πœƒ−1) = 2 whenever 𝑑 ∈ (πœƒ − 1, πœƒ + 1) and
𝑑
1
1
𝐹𝑋 (𝑑) = ∫πœƒ−1 2 𝑑π‘₯ = 2 [π‘₯]π‘‘πœƒ−1 =
𝑑−(πœƒ−1)
2
𝑑 ∈ (πœƒ − 1, πœƒ + 1).
whenever
Then
the
distribution function of 𝑋(𝑛) is given by 𝐹𝑛 (𝑑) = 𝑃(𝑋(𝑛) ≤ 𝑑) = 𝑃(𝑋𝑖 ≤ 𝑑)𝑛 =
(
𝑑−(πœƒ−1) 𝑛
(𝑑−πœƒ+1)𝑛
) =
2
𝑑
so the density function of 𝑋(𝑛) is 𝑓𝑛 (𝑑) = 𝑑𝑑 𝐹𝑛 (𝑑) =
2𝑛
We can then compute the mean of 𝑋(𝑛)
πœƒ+1
∫πœƒ−1 𝑑
𝑛(𝑑−πœƒ+1)𝑛−1
2𝑛
𝑒 =𝑑−πœƒ+1
πœƒ+1
∫πœƒ−1 𝑑
𝑛
𝑒𝑛+1
2𝑛
[
+
2𝑛 𝑛+1
πœƒπ‘’π‘›
𝑛
−
2𝑛
.
πœƒ+1
as 𝐸(𝑋(𝑛) ) = ∫πœƒ−1 𝑑𝑓𝑛 (𝑑) 𝑑𝑑 =
𝑑𝑑. This integral can be calculated by completing the substitution
so
𝑛(𝑑−πœƒ+1)𝑛−1
𝑛(𝑑−πœƒ+1)𝑛−1
that
𝑑𝑒 = 𝑑𝑑
𝑛
2
𝑛
2𝑛+1
and
𝑑 = 𝑒 + πœƒ − 1.
𝑛
This
then
implies
2
𝑑𝑑 = 2𝑛 ∫0 (𝑒 + πœƒ − 1)𝑒𝑛−1 𝑑𝑒 = 2𝑛 ∫0 𝑒𝑛 + πœƒπ‘’π‘›−1 − 𝑒𝑛−1 𝑑𝑒 =
𝑒𝑛
2
] = 2𝑛 [ 𝑛+1 +
𝑛
πœƒ2𝑛
𝑛
0
−
2𝑛
2𝑛
] = 𝑛+1 + πœƒ − 1. We can similarly compute
𝑛
2𝑛
that the expected value of the first order statistic is 𝐸(𝑋(1) ) = − 𝑛+1 + πœƒ + 1. Thus, we
have
1
2
1
that
1
2𝑛
2𝑛
𝐸(𝑀) = 2 [𝐸(𝑋(1) ) + 𝐸(𝑋(𝑛) )] = 2 [(πœƒ − 𝑛+1 + 1) + (πœƒ + 𝑛+1 − 1)] =
[2πœƒ] = πœƒ, so the midrange is also an unbiased estimator for the parameter πœƒ.
c) We
1
𝑛2
have
∑𝑛𝑖=1
that
1
1
1
π‘‰π‘Žπ‘Ÿ(𝑋̅) = π‘‰π‘Žπ‘Ÿ (𝑛 ∑𝑛𝑖=1 𝑋𝑖 ) = 𝑛2 π‘‰π‘Žπ‘Ÿ(∑𝑛𝑖=1 𝑋𝑖 ) = 𝑛2 ∑𝑛𝑖=1 π‘‰π‘Žπ‘Ÿ(𝑋𝑖 ) =
[(πœƒ+1)−(πœƒ−1)]2
12
1
4
1
1
1 𝑛
1
= 𝑛2 ∑𝑛𝑖=1 12 = 𝑛2 ∑𝑛𝑖=1 3 = 𝑛2 3 = 3𝑛. Similarly, we can calculate
𝑋(1) +𝑋(𝑛)
that π‘‰π‘Žπ‘Ÿ(𝑀) = π‘‰π‘Žπ‘Ÿ (
2
1
) = 4 π‘‰π‘Žπ‘Ÿ(𝑋(1) + 𝑋(𝑛) ).
Question #21: Let 𝑋1 , … , 𝑋𝑛 be a random sample from 𝑋~𝐡𝐼𝑁(1, 𝑝). a) Find the Cramer-Rao
lower bound for the variances of all unbiased estimators of 𝑝; b) Find the Cramer-Rao lower
bound for the variances of unbiased estimators of 𝑝(1 − 𝑝); c) Find a UMVUE of 𝑝.
2
a) We have that 𝐢𝑅𝐿𝐡 =
[𝜏′ (𝑝)]
𝑛𝐸[(
, so we compute each of these parts individually.
2
πœ•
ln 𝑓(𝑋;𝑝)) ]
πœ•π‘
First, we have 𝜏(𝑝) = 𝑝, so 𝜏 ′ (𝑝) = 1 and [𝜏 ′ (𝑝)]2 = 1. Next, since 𝑋~𝐡𝐼𝑁(1, 𝑝) we
know that 𝑓𝑋 (π‘₯) = (π‘₯1)𝑝 π‘₯ (1 − 𝑝)1−π‘₯ = 𝑝 π‘₯ (1 − 𝑝)1−π‘₯ and so 𝑓(𝑋; 𝑝) = 𝑝 𝑋 (1 − 𝑝)1−𝑋 ,
which means that ln 𝑓(𝑋; 𝑝) = 𝑋 ln(𝑝) + (1 − 𝑋) ln(1 − 𝑝). Taking the derivative and
squaring
πœ•
gives
πœ•π‘
2
πœ•
𝑋
𝑋−𝑝
2
(πœ•π‘ ln 𝑓(𝑋; 𝑝)) = (𝑝(1−𝑝)) =
𝐸[
𝑋 2 −2𝑝𝑋+𝑝2
]
𝑝2 (1−𝑝)2
1
𝑝2 (1−𝑝)2
1
𝑝2 (1−𝑝)2
1−𝑋
ln 𝑓(𝑋; 𝑝) = 𝑝 − 1−𝑝 =
𝑋 2 −2𝑝𝑋+𝑝2
.
𝑝2 (1−𝑝)2
(1−𝑝)𝑋−𝑝(1−𝑋)
𝑝(1−𝑝)
=
𝑋−𝑝𝑋−𝑝+𝑝𝑋
𝑝(1−𝑝)
𝑋−𝑝
= 𝑝(1−𝑝) →
2
πœ•
Finally, we compute 𝐸 [(πœ•π‘ ln 𝑓(𝑋; 𝑝)) ] =
1
1
= 𝑝2 (1−𝑝)2 𝐸(𝑋 2 − 2𝑝𝑋 + 𝑝2 ) = 𝑝2 (1−𝑝)2 [𝐸(𝑋 2 ) − 2𝑝𝐸(𝑋) + 𝑝2 ] =
[(𝑝(1 − 𝑝) + 𝑝2 ) − 2𝑝(𝑝) + 𝑝2 ] =
[𝑝 − 𝑝2 ] =
𝑝(1−𝑝)
𝑝2 (1−𝑝)2
1
𝑝2 (1−𝑝)2
[𝑝 − 𝑝2 + 𝑝2 − 2𝑝2 + 𝑝2 ] =
1
= 𝑝(1−𝑝). Thus, we have found that 𝐢𝑅𝐿𝐡 =
𝑝(1−𝑝)
𝑛
.
b) Now, 𝜏(𝑝) = 𝑝(1 − 𝑝) = 𝑝 − 𝑝2 , so [𝜏 ′ (𝑝)]2 = [1 − 2𝑝]2 = 1 − 4𝑝 + 4𝑝2 , so the
Cramer-Rao Lower Bound becomes 𝐢𝑅𝐿𝐡 =
(1−4𝑝+4𝑝2 )𝑝(1−𝑝)
𝑛
.
1
1
c) Since for the estimator 𝑝̂ = 𝑋̅, we have 𝐸(𝑝̂ ) = 𝐸(𝑋̅) = 𝐸 (𝑛 ∑𝑛𝑖=1 𝑋𝑖 ) = 𝑛 𝐸(∑𝑛𝑖=1 𝑋𝑖 ) =
1
𝑛
1
1
1
∑𝑛𝑖=1 𝐸(𝑋𝑖 ) = ∑𝑛𝑖=1 𝑝 = 𝑛𝑝 = 𝑝 and then π‘‰π‘Žπ‘Ÿ(𝑝̂ ) = π‘‰π‘Žπ‘Ÿ(𝑋̅) = π‘‰π‘Žπ‘Ÿ ( ∑𝑛𝑖=1 𝑋𝑖 ) =
𝑛
𝑛
𝑛
1
𝑛2
1
1
1
π‘‰π‘Žπ‘Ÿ(∑𝑛𝑖=1 𝑋𝑖 ) = 𝑛2 ∑𝑛𝑖=1 π‘‰π‘Žπ‘Ÿ(𝑋𝑖 ) = 𝑛2 ∑𝑛𝑖=1 𝑝(1 − 𝑝) = 𝑛2 𝑛𝑝(1 − 𝑝) =
𝑝(1−𝑝)
𝑛
=
𝐢𝑅𝐿𝐡, we can conclude that 𝑝̂ = 𝑋̅ is a Uniform Minimum Variance Unbiased
Estimator (UMVUE) for the parameter 𝑝 in 𝑋~𝐡𝐼𝑁(1, 𝑝).
Question #22: Let 𝑋1 , … , 𝑋𝑛 be a random sample from 𝑋~𝑁(πœ‡, 9). a) Find the Cramer-Rao
lower bound for the variances of unbiased estimators of πœ‡; b) is the Maximum Likelihood
Estimator πœ‡Μ‚ 𝑀𝐿𝐸 = 𝑋̅ a UMVUE for the parameter πœ‡?
a) We have 𝜏(πœ‡) = πœ‡, so 𝜏 ′ (πœ‡) = 1 and [𝜏 ′ (πœ‡)]2 = 1. Next, since 𝑋~𝑁(πœ‡, 9) we know that
the density is 𝑓𝑋 (π‘₯) =
1
1
√18πœ‹
1
𝑒 −18
(π‘₯−πœ‡)2
, so that we have 𝑓(𝑋; πœ‡) =
1
√18πœ‹
1
𝑒 −18
(𝑋−πœ‡)2
and
1
ln 𝑓(𝑋; πœ‡) = − 2 ln(18πœ‹) − 18 (𝑋 − πœ‡)2. We then differentiate twice to obtain
πœ•
πœ•πœ‡
1
πœ•2
9
πœ•πœ‡ 2
ln 𝑓(𝑋; πœ‡) = (𝑋 − πœ‡) →
1
ln 𝑓(𝑋; πœ‡) = − . Since we have shown than the
2
πœ•
9
πœ•2
1
1
expression 𝐸 [(πœ•πœ‡ ln 𝑓(𝑋; 𝑝)) ] = −𝐸 [πœ•πœ‡2 ln 𝑓(𝑋; πœ‡)] and −𝐸 (− 9) = 9, we can
9
conclude that the Cramer-Rao Lower Bound is 𝐢𝑅𝐿𝐡 = 𝑛. This then means that
9
π‘‰π‘Žπ‘Ÿ(𝑇) ≥ 𝑛 for any unbiased estimator 𝑇 of the parameter πœ‡ in 𝑋~𝑁(πœ‡, 9).
1
1
1
b) We first verify that 𝐸(πœ‡Μ‚ 𝑀𝐿𝐸 ) = 𝐸(𝑋̅) = 𝐸 (𝑛 ∑𝑛𝑖=1 𝑋𝑖 ) = 𝑛 𝐸(∑𝑛𝑖=1 𝑋𝑖 ) = 𝑛 ∑𝑛𝑖=1 𝐸(𝑋𝑖 ) =
1
1
∑𝑛 πœ‡ = π‘›πœ‡ = πœ‡, so πœ‡Μ‚ 𝑀𝐿𝐸 = 𝑋̅ is an unbiased estimator for πœ‡. Then we compute
𝑛 𝑖=1
𝑛
1
1
1
π‘‰π‘Žπ‘Ÿ(πœ‡Μ‚ 𝑀𝐿𝐸 ) = π‘‰π‘Žπ‘Ÿ(𝑋̅) = π‘‰π‘Žπ‘Ÿ (𝑛 ∑𝑛𝑖=1 𝑋𝑖 ) = 𝑛2 π‘‰π‘Žπ‘Ÿ(∑𝑛𝑖=1 𝑋𝑖 ) = 𝑛2 ∑𝑛𝑖=1 π‘‰π‘Žπ‘Ÿ(𝑋𝑖 ) =
1
𝑛2
∑𝑛𝑖=1 9 =
1
𝑛2
9
9𝑛 = 𝑛 = 𝐢𝑅𝐿𝐡, so that πœ‡Μ‚ 𝑀𝐿𝐸 = 𝑋̅ a UMVUE for the parameter πœ‡.
Question #23: Let 𝑋1 , … , 𝑋𝑛 be a random sample from 𝑋~𝑁(0, πœƒ). a) Is the Maximum
Likelihood Estimator (MLE) for πœƒ unbiased?; b) is the MLE also a UMVUE for πœƒ?
a) We first find πœƒΜ‚π‘€πΏπΈ by noting that since 𝑋~𝑁(0, πœƒ), then its density function is 𝑓𝑋 (π‘₯) =
1
√2πœ‹πœƒ
1
2
𝑒 −2πœƒπ‘₯ so the likelihood function is 𝐿(πœƒ) = ∏𝑛𝑖=1 𝑓𝑋 (π‘₯; πœƒ) = ∏𝑛𝑖=1
𝑛
−
2
(2πœ‹πœƒ) 𝑒
1
− ∑𝑛
π‘₯2
2πœƒ 𝑖=1 𝑖
differentiate
so
and
that
1
then
πœ•
πœ•πœƒ
𝑛
1
ln[𝐿(πœƒ)] = − 2 ln(2πœ‹πœƒ) − 2πœƒ ∑𝑛𝑖=1 π‘₯𝑖2 .
𝑛2πœ‹
1
1
1
√2πœ‹πœƒ
1
2
𝑒 −2πœƒπ‘₯𝑖 =
Next,
we
𝑛
ln[𝐿(πœƒ)] = − 4πœ‹πœƒ + 2πœƒ2 ∑𝑛𝑖=1 π‘₯𝑖2 = 0 → 2πœƒ2 ∑𝑛𝑖=1 π‘₯𝑖2 = 2πœƒ →
∑𝑛𝑖=1 π‘₯𝑖2 = π‘›πœƒ → πœƒ = ∑𝑛𝑖=1 π‘₯𝑖2 . Since the second derivative is negative, we have
𝑛
1
1
πœƒΜ‚π‘€πΏπΈ = 𝑛 ∑𝑛𝑖=1 𝑋𝑖2 . We verify unbiasedness by computing 𝐸(πœƒΜ‚π‘€πΏπΈ ) = 𝐸 (𝑛 ∑𝑛𝑖=1 𝑋𝑖2 ) =
1
1
1
1
1
𝐸(∑𝑛𝑖=1 𝑋𝑖2 ) = 𝑛 ∑𝑛𝑖=1 𝐸(𝑋𝑖2 ) = 𝑛 ∑𝑛𝑖=1(πœƒ + 02 ) = 𝑛 ∑𝑛𝑖=1 πœƒ = 𝑛 π‘›πœƒ = πœƒ.
𝑛
b) The estimator πœƒΜ‚π‘€πΏπΈ will be a UMVUE for πœƒ if π‘‰π‘Žπ‘Ÿ(πœƒΜ‚π‘€πΏπΈ ) = 𝐢𝑅𝐿𝐡. We therefore begin
by computing the Cramer-Rao Lower Bound. First, we have 𝜏(πœƒ) = πœƒ, so 𝜏 ′ (πœƒ) = 1
and [𝜏 ′ (πœƒ)]2 = 1. Next, since we previously found that 𝑓(𝑋; πœƒ) =
1
1
πœ•
1
1
√2πœ‹πœƒ
2
𝑒 −2πœƒπ‘‹ , then we
𝑋2
1
have ln 𝑓(𝑋; πœƒ) = − 2 ln(2πœ‹πœƒ) − 2πœƒ 𝑋 2 so that πœ•πœƒ ln 𝑓(𝑋; πœƒ) = − 2πœƒ + 2πœƒ2. We then find
πœ•2
2𝑋 2
1
1
𝑋2
ln 𝑓(𝑋; πœƒ) = 2πœƒ2 − 2πœƒ3 = 2πœƒ2 − πœƒ3 and take the negative of its expected value to
πœ•πœƒ2
1
𝑋2
1
1
1
1
1
1
1
obtain −𝐸 [2πœƒ2 − πœƒ3 ] = − [2πœƒ2 − πœƒ3 𝐸(𝑋 2 )] = πœƒ3 (πœƒ + 02 ) − 2πœƒ2 = πœƒ2 − 2πœƒ2 = 2πœƒ2. This
implies that 𝐢𝑅𝐿𝐡 =
to
1
𝑛2
this
lower
2πœƒ2
𝑛
. We must verify that the variance of our estimator is equal
bound,
so
we
compute
1
π‘‰π‘Žπ‘Ÿ(πœƒΜ‚π‘€πΏπΈ ) = π‘‰π‘Žπ‘Ÿ (𝑛 ∑𝑛𝑖=1 𝑋𝑖2 ) =
1
π‘‰π‘Žπ‘Ÿ(∑𝑛𝑖=1 𝑋𝑖2 ) = 𝑛2 ∑𝑛𝑖=1 π‘‰π‘Žπ‘Ÿ(𝑋𝑖2 ). In order to compute π‘‰π‘Žπ‘Ÿ(𝑋𝑖2 ), we use the formula
π‘‰π‘Žπ‘Ÿ(𝑋𝑖2 ) = 𝐸(𝑋𝑖4 ) − [𝐸(𝑋𝑖2 )]2 = β‹― = 3πœƒ 2 − πœƒ 2 = 2πœƒ 2 by finding the moments of 𝑋𝑖
using the derivatives of the Moment Generating Function at 𝑑 = 0. Then we have that
2
1
1
1
2πœƒ
π‘‰π‘Žπ‘Ÿ(πœƒΜ‚π‘€πΏπΈ ) = 𝑛2 ∑𝑛𝑖=1 π‘‰π‘Žπ‘Ÿ(𝑋𝑖2 ) = 𝑛2 ∑𝑛𝑖=1 2πœƒ 2 = 𝑛2 𝑛2πœƒ 2 = 𝑛 = 𝐢𝑅𝐿𝐡, which verifies
that the Maximum Likelihood Estimator is a UMVUE for the parameter πœƒ.
Chapter #9 – Point Estimation
Question #31: Let πœƒΜ‚ and πœƒΜƒ be the MLE and MME estimators for the parameter πœƒ, where
𝑋1 , … , 𝑋𝑛 is a random sample of size 𝑛 from a Uniform distribution such that 𝑋𝑖 ~π‘ˆπ‘πΌπΉ(0, πœƒ).
Show that a) πœƒΜ‚ is MSE consistent, and b) πœƒΜƒ is MSE consistent.
a) We first derive the MLE πœƒΜ‚ for πœƒ. Since 𝑋~π‘ˆπ‘πΌπΉ(0, πœƒ), we know that the density
1
function is 𝑓(π‘₯; πœƒ) = πœƒ for π‘₯ ∈ (0, πœƒ). This allows us to construct the likelihood
function 𝐿(πœƒ) = ∏𝑛𝑖=1 𝑓(π‘₯𝑖 ; πœƒ) = ∏𝑛𝑖=1 πœƒ −1 = πœƒ −𝑛 whenever π‘₯1:𝑛 ≥ 0 and π‘₯𝑛:𝑛 ≤ πœƒ and
zero otherwise. Then the log likelihood function is ln[𝐿(πœƒ)] = −𝑛 ln(πœƒ) so that
πœ•
πœ•πœƒ
𝑛
ln[𝐿(πœƒ)] = − πœƒ < 0 for all 𝑛 and πœƒ. This means that 𝐿(πœƒ) = πœƒ −𝑛 is a decreasing
function of πœƒ for π‘₯𝑛:𝑛 ≤ πœƒ since its first derivative is always negative, so we can
conclude that the MLE is the largest order statistic, so πœƒΜ‚ = 𝑋𝑛:𝑛 . Next, we show that
this estimator is MSE consistent, which means verifying that lim 𝐸[𝑋𝑛:𝑛 − πœƒ]2 = 0.
𝑛→∞
But then we can see that
2
lim [𝐸(𝑋𝑛:𝑛
)
𝑛→∞
− 2πœƒπΈ(𝑋𝑛:𝑛 ) + πœƒ
2 ].
2
lim 𝐸[𝑋𝑛:𝑛 − πœƒ]2 = lim 𝐸[𝑋𝑛:𝑛
− 2πœƒπ‘‹π‘›:𝑛 + πœƒ 2 ] =
𝑛→∞
𝑛→∞
In order to compute this limit, we must find the first
and second moments of the largest order statistic. But we already know that 𝑓𝑛 (𝑦) =
𝑛 𝑦 𝑛−1
𝑛𝑓(𝑦; πœƒ)𝐹(𝑦; πœƒ)𝑛−1 = πœƒ (πœƒ)
πœƒ
∫0 𝑦
𝑛𝑦 𝑛−1
πœƒπ‘›
πœƒ
𝑛
=
𝑛
𝑛𝑦 𝑛−1
πœƒπ‘›
𝑦 𝑛+1
πœƒ
, so we can calculate 𝐸(𝑋𝑛:𝑛 ) = ∫0 𝑦𝑓𝑛 (𝑦) 𝑑𝑦 =
πœƒ
𝑛
πœƒπ‘›+1
𝑛
𝑑𝑦 = πœƒπ‘› ∫0 𝑦 𝑛 𝑑𝑦 = πœƒπ‘› [ 𝑛+1 ] = πœƒπ‘› ( 𝑛+1 − 0) = 𝑛+1 πœƒ
0
πœƒ
πœƒ
∫0 𝑦 2 𝑓𝑛 (𝑦) 𝑑𝑦 = ∫0 𝑦 2
𝑛𝑦 𝑛−1
πœƒπ‘›
𝑛
πœƒ
𝑛
𝑦 𝑛+2
πœƒ
𝑛
2 )
𝐸(𝑋𝑛:𝑛
=
and
πœƒπ‘›+2
𝑛
𝑑𝑦 = πœƒπ‘› ∫0 𝑦 𝑛+1 𝑑𝑦 = πœƒπ‘› [ 𝑛+2 ] = πœƒπ‘› ( 𝑛+2 ) = 𝑛+2 πœƒ 2 .
0
𝑛
𝑛
2
Thus, we have lim [𝐸(𝑋𝑛:𝑛
) − 2πœƒπΈ(𝑋𝑛:𝑛 ) + πœƒ 2 ] = lim [𝑛+2 πœƒ 2 − 2πœƒ 𝑛+1 πœƒ + πœƒ 2 ] =
𝑛→∞
lim [
𝑛
𝑛→∞ 𝑛+2
𝑛→∞
2𝑛
πœƒ 2 − 𝑛+1 πœƒ 2 + πœƒ 2 ] = πœƒ 2 − 2πœƒ 2 + πœƒ 2 = 0. That this limit is zero verifies that
the maximum likelihood estimator πœƒΜ‚ = 𝑋𝑛:𝑛 is mean square error (MSE) consistent.
πœƒ
b) We first derive the MME πœƒΜƒ for πœƒ. Since 𝑋~π‘ˆπ‘πΌπΉ(0, πœƒ), we know that 𝐸(𝑋) = 2 so we
can equate πœ‡1′ = 𝑀1′ →
πœƒ
2
1
= 𝑛 ∑𝑛𝑖=1 𝑋𝑖 →
πœƒ
2
= 𝑋̅ → πœƒ = 2𝑋̅. This means that πœƒΜƒ = 2𝑋̅.
Next, we show that this estimator is MSE consistent, which means verifying that
lim 𝐸[2𝑋̅ − πœƒ]2 = 0. But we have lim 𝐸[2𝑋̅ − πœƒ]2 = lim 𝐸[4𝑋̅ 2 − 4πœƒπ‘‹Μ… + πœƒ 2 ] =
𝑛→∞
𝑛→∞
lim [4𝐸(𝑋̅ 2 ) − 4πœƒπΈ(𝑋̅) + πœƒ 2 ].
We
𝑛→∞
1
𝑛
1
1
𝑛→∞
therefore
πœƒ
1
πœƒ
𝐸(∑𝑛𝑖=1 𝑋𝑖 ) = 𝑛 ∑𝑛𝑖=1 𝐸(𝑋𝑖 ) = 𝑛 ∑𝑛𝑖=1 2 = 𝑛 𝑛 2 =
πœƒ 2
1
1
π‘‰π‘Žπ‘Ÿ (𝑛 ∑𝑛𝑖=1 𝑋𝑖 ) + (2) = 𝑛2 ∑𝑛𝑖=1 π‘‰π‘Žπ‘Ÿ(𝑋𝑖 ) +
πœƒ2
+
12𝑛
πœƒ2
4
lim [4
𝑛→∞
=
(3𝑛+1)πœƒ 2
12𝑛
(3𝑛+1)πœƒ2
12𝑛
πœƒ2
1
𝐸(𝑋̅) = 𝐸 (𝑛 ∑𝑛𝑖=1 𝑋𝑖 ) =
compute
πœƒ
2
and 𝐸(𝑋̅ 2 ) = π‘‰π‘Žπ‘Ÿ(𝑋̅) + 𝐸(𝑋̅)2 =
πœƒ2
1
= 𝑛2 ∑𝑛𝑖=1 12 +
4
πœƒ2
4
1
πœƒ2
= 𝑛2 𝑛 12 +
πœƒ2
4
=
. Thus, we can compute that lim [4𝐸(𝑋̅ 2 ) − 4πœƒπΈ(𝑋̅) + πœƒ 2 ] =
𝑛→∞
πœƒ
− 4πœƒ + πœƒ 2 ] = lim [
2
(3𝑛+1)πœƒ2
3𝑛
𝑛→∞
− 2πœƒ 2 + πœƒ 2 ] = πœƒ 2 − 2πœƒ 2 + πœƒ 2 = 0. That
this limit is zero verifies that the MME πœƒΜƒ = 2𝑋̅ is mean square error (MSE) consistent.
Question #29: Let 𝑋1 , … , 𝑋𝑛 be a random sample of size 𝑛 from a Bernoulli distribution such
that 𝑋𝑖 ~𝐡𝐼𝑁(1, 𝑝). For a Uniform prior density 𝑝~π‘ˆπ‘πΌπΉ(0,1) and a squared error loss
function 𝐿(𝑑; 𝑝) = (𝑑 − 𝑝)2 , a) find the posterior distribution of the unknown parameter 𝑝,
b) find the Bayes estimator of 𝑝, and c) find the Bayes risk for the Bayes estimator of 𝑝 above.
a) We have that the posterior density is given by 𝑓P|π‘₯ (𝑝) =
𝑓(π‘₯1 ,…,π‘₯𝑛 ;𝑝)𝑝(𝑝)
∫ 𝑓(π‘₯1 ,…,π‘₯𝑛 ;𝑝)𝑝(𝑝)𝑑𝑝
𝑛
, where
𝑛
𝑓(π‘₯1 , … , π‘₯𝑛 ; 𝑝) = ∏𝑛𝑖=1 𝑓(π‘₯𝑖 ; 𝑝) = ∏𝑛𝑖=1 𝑝 π‘₯ (1 − 𝑝)1−π‘₯ = 𝑝∑𝑖=1 π‘₯𝑖 (1 − 𝑝)𝑛−∑𝑖=1 π‘₯𝑖
since
the random variables are independent and identically distributed and 𝑝(𝑝) = 1 since
1
𝑛
𝑛
the prior density is uniform. We then express ∫0 𝑝∑𝑖=1 π‘₯𝑖 (1 − 𝑝)𝑛−∑𝑖=1 π‘₯𝑖 𝑑𝑝 in terms of
the beta distribution. Recall that if π‘Œ~𝐡𝐸𝑇𝐴(π‘Ž, 𝑏), then its density is 𝑓(𝑦; π‘Ž, 𝑏) =
1
𝑦 π‘Ž−1 (1 − 𝑦)𝑏−1 where 𝐡(π‘Ž, 𝑏) =
𝐡(π‘Ž,𝑏)
1
Γ(π‘Ž)Γ(𝑏)
Γ(π‘Ž+𝑏)
𝑛
. Next, we must define π‘Ž = ∑𝑛𝑖=1 π‘₯𝑖 and
𝑛
𝑏 = 𝑛 − ∑𝑛𝑖=1 π‘₯𝑖 , so we can write ∫0 𝑝∑𝑖=1 π‘₯𝑖 (1 − 𝑝)𝑛−∑𝑖=1 π‘₯𝑖 𝑑𝑝 = 𝐡(π‘Ž + 1, 𝑏 + 1) =
𝑛
𝑛
𝑝∑𝑖=1 π‘₯𝑖 (1−𝑝)𝑛−∑𝑖=1 π‘₯𝑖
𝐡(∑𝑛𝑖=1 π‘₯𝑖 + 1, 𝑛 − ∑𝑛𝑖=1 π‘₯𝑖 + 1). Thus, we have 𝑓P|π‘₯ (𝑝) = 𝐡(∑𝑛
𝑛
𝑖=1 π‘₯𝑖 +1,𝑛−∑𝑖=1 π‘₯𝑖 +1)
1
𝐡(π‘Ž+1,𝑏+1)
=
π‘π‘Ž (1 − 𝑝)𝑏 , which verifies that the random variable given by
P|π‘₯~𝐡𝐸𝑇𝐴(∑𝑛𝑖=1 π‘₯𝑖 + 1, 𝑛 − ∑𝑛𝑖=1 π‘₯𝑖 + 1) ≡ 𝐡𝐸𝑇𝐴(π‘Ž + 1, 𝑏 + 1).
π‘Ž
b) For some random variable π‘Œ~𝐡𝐸𝑇𝐴(π‘Ž, 𝑏), we know that 𝐸(π‘Œ) = π‘Ž+𝑏. Moreover,
Theorem 9.5.2 states that when we have a squared error loss function, the Bayes
estimator is simply the expected value of the posterior distribution. This implies that
∑𝑛
𝑖=1 π‘₯𝑖 +1
the Bayes estimator of 𝑝 is given by 𝑝̂ 𝐡𝐸 = ∑𝑛
𝑛
𝑖=1 π‘₯𝑖 +1+𝑛−∑𝑖=1 π‘₯𝑖 +1
=
c) The risk function in this case is 𝑅𝑇 (𝑝) = 𝐸[(𝑇 − 𝑝)2 ], where 𝑇 =
∑𝑛
𝑖=1 π‘₯𝑖 +1
𝑛+2
.
∑𝑛
𝑖=1 π‘₯𝑖 +1
𝑛+2
is the Bayes
Estimator derived above. We would therefore substitute for 𝑇 in the risk function,
1
evaluate the expected value of that expression and then compute ∫0 𝐸[(𝑇 − 𝑝)2 ] 𝑑𝑝.
Question #34: Consider a random sample of size 𝑛 from a distribution with discrete
probability mass function 𝑓𝑋 (π‘₯; 𝑝) = (1 − 𝑝)π‘₯ 𝑝 for π‘₯ ∈ {0,1,2, … }. a) Find the MLE of the
unknown parameter 𝑝. b) Find the MLE of πœƒ =
1−𝑝
𝑝
. c) Find the CRLB for variances of all
unbiased estimators of the parameter πœƒ above. d) Is the MLE of πœƒ =
MLE of πœƒ =
πœƒ=
1−𝑝
𝑝
1−𝑝
𝑝
1−𝑝
𝑝
a UMVUE? e) Is the
also MSE consistent? f) Compute the asymptotic distribution of the MLE of
𝑛
. g) If we have the estimator πœƒΜ‚ = 𝑛+1 𝑋̅, then find the risk functions of both πœƒΜ‚ and 𝑋̅
using the loss function given by 𝐿(𝑑; πœƒ) =
a) We
have
(𝑑−πœƒ)2
πœƒ2 +πœƒ
.
𝑛
𝐿(𝑝) = ∏𝑝𝑖=1 𝑓(π‘₯𝑖 ; 𝑝) = ∏𝑝𝑖=1(1 − 𝑝)π‘₯𝑖 𝑝 = 𝑝𝑛 (1 − 𝑝)∑𝑖=1 π‘₯𝑖 ,
ln[𝐿(𝑝)] = 𝑛 ln(𝑝) + ∑𝑛𝑖=1 π‘₯𝑖 ln(1 − 𝑝). Then we have that
πœ•
so
𝑛
ln[𝐿(𝑝)] = 𝑝 −
πœ•π‘
that
∑𝑛
𝑖=1 π‘₯𝑖
1−𝑝
.
1
Setting this equal to zero and solving for 𝑝 gives the estimator 𝑝̂ 𝑀𝐿𝐸 = 1+𝑋̅.
1−𝑝̂
b) By the Invariance Property, we have that the estimator is πœƒΜ‚π‘€πΏπΈ = 𝑝̂ 𝑀𝐿𝐸 = 𝑋̅.
𝑀𝐿𝐸
c) Since πœƒ = 𝜏(𝑝) =
1−𝑝
𝑝
1
1
1
= 𝑝 − 1, then 𝜏 ′ (𝑝) = − 𝑝2 and [𝜏 ′ (𝑝)]2 = 𝑝4. Then since
𝑓(𝑋; 𝑝) = (1 − 𝑝) 𝑋 𝑝, we can compute ln 𝑓(𝑋; 𝑝) = ln(𝑝) + 𝑋ln(1 − 𝑝) so that
πœ•
1
πœ•2
𝑋
ln 𝑓(𝑋; 𝑝) = 𝑝 − 1−𝑝 and
πœ•π‘
πœ•π‘2
1
𝑋
ln 𝑓(𝑋; 𝑝) = − 𝑝2 − (1−𝑝)2. We can then compute the
πœ•2
negative of the expected value of this second derivative so that −𝐸 [πœ•π‘2 ln 𝑓(𝑋; 𝑝)] =
1
1
1
1−𝑝
+ (1−𝑝)2 𝐸(𝑋) = 𝑝2 + 𝑝(1−𝑝)2 =
𝑝2
These results imply that 𝐢𝑅𝐿𝐡 =
(1−𝑝)2 +𝑝(1−𝑝)
=
𝑝2 (1−𝑝)2
𝑝2 (1−𝑝)
𝑛𝑝4
=
1−𝑝
𝑛𝑝2
1−2𝑝+𝑝2 +𝑝−𝑝2
𝑝2 (1−𝑝)2
1−𝑝
1
= 𝑝2 (1−𝑝)2 = 𝑝2 (1−𝑝).
.
1
1
1 1−𝑝
1−𝑝
d) We first verify that 𝐸(πœƒΜ‚π‘€πΏπΈ ) = 𝐸(𝑋̅) = 𝐸 (𝑛 ∑𝑛𝑖=1 𝑋𝑖 ) = 𝑛 ∑𝑛𝑖=1 𝐸(𝑋𝑖 ) = 𝑛 𝑛 𝑝 = 𝑝 ,
so that the MLE is an unbiased estimator of πœƒ =
1−𝑝
𝑝
. Next, we compute π‘‰π‘Žπ‘Ÿ(πœƒΜ‚π‘€πΏπΈ ) =
1
1
1
1−𝑝
1−𝑝
π‘‰π‘Žπ‘Ÿ(𝑋̅) = π‘‰π‘Žπ‘Ÿ (𝑛 ∑𝑛𝑖=1 𝑋𝑖 ) = 𝑛2 ∑𝑛𝑖=1 π‘‰π‘Žπ‘Ÿ(𝑋𝑖 ) = 𝑛2 𝑛 𝑝2 = 𝑛𝑝2 = 𝐢𝑅𝐿𝐡, which verifies
that πœƒΜ‚π‘€πΏπΈ = 𝑋̅ is the UMVUE for the parameter πœƒ =
1−𝑝
𝑝
.
2
1−𝑝
e) To verify that πœƒΜ‚π‘€πΏπΈ = 𝑋̅ is MSE consistent, we must show that lim 𝐸 [𝑋̅ − 𝑝 ] = 0.
𝑛→∞
2
2
(1−𝑝)
1−𝑝
2(1−𝑝)
But we can see that we have lim 𝐸 [𝑋̅ − 𝑝 ] = lim 𝐸 [𝑋̅ 2 − 𝑝 𝑋̅ + 𝑝2 ] =
𝑛→∞
2(1−𝑝)
lim [𝐸(𝑋̅ 2 ) − 𝑝 𝐸(𝑋̅) +
(1−𝑝)2
𝑛→∞
𝑝2
𝑛→∞
], so we must compute the expectation of both the
1−𝑝
mean and the mean squared. However, we already know that 𝐸(𝑋̅) = 𝑝 = πœƒ since
2
1
1−𝑝
πœƒΜ‚π‘€πΏπΈ is unbiased. Then 𝐸(𝑋̅ 2 ) = π‘‰π‘Žπ‘Ÿ(𝑋̅) + 𝐸(𝑋̅)2 = π‘‰π‘Žπ‘Ÿ (𝑛 ∑𝑛𝑖=1 𝑋𝑖 ) + ( 𝑝 ) =
1
𝑛2
∑𝑛𝑖=1 π‘‰π‘Žπ‘Ÿ(𝑋𝑖 ) +
(1−𝑝)2
𝑝2
1
= 𝑛2 ∑𝑛𝑖=1
1−𝑝
𝑝2
+
(1−𝑝)2
𝑝2
1
= 𝑛2 𝑛
1−𝑝
𝑝2
+
2
(1−𝑝)2
𝑝2
=
1−𝑝
𝑛𝑝2
+
2
(1−𝑝)2
𝑝2
.
2
(1−𝑝)
(1−𝑝)
(1−𝑝)
2(1−𝑝)
1−𝑝
2(1−𝑝) 1−𝑝
Thus lim [𝐸(𝑋̅ 2 ) − 𝑝 𝐸(𝑋̅) + 𝑝2 ] = lim [ 𝑛𝑝2 + 𝑝2 − 𝑝
+
]=
𝑝
𝑝2
𝑛→∞
(1−𝑝)2
𝑝2
−
𝑛→∞
2(1−𝑝)2
𝑝2
+
(1−𝑝)2
𝑝2
= 0. This shows that πœƒΜ‚π‘€πΏπΈ = 𝑋̅ is MSE consistent.
f) We use Definition 9.4.5, which states that for large values of 𝑛, the MLE estimator is
distributed normal with mean πœƒ =
that 𝐢𝑅𝐿𝐡 =
1−𝑝
𝑛𝑝2
1−𝑝
𝑝
and variance 𝐢𝑅𝐿𝐡. Since we previously found
1−𝑝 1−𝑝
, we can conclude that πœƒΜ‚π‘€πΏπΈ ~𝑁 ( 𝑝 , 𝑛𝑝2 ).
g) Definition 9.5.2 states that the risk function is the expected loss 𝑅𝑇 (πœƒ) = 𝐸[𝐿(𝑇; πœƒ)].
In this case, the loss function is 𝐿(𝑑; πœƒ) =
(𝑑−πœƒ)2
πœƒ2 +πœƒ
Μ… 2 −2πœƒπ‘‹Μ… +πœƒ2
𝑋
estimator 𝑋̅ we compute 𝑅𝑋̅ (πœƒ) = 𝐸 [
1
πœƒ2 +πœƒ
1−𝑝
[ 𝑛𝑝2 +
πœƒ
(πœƒ+1)𝑛𝑝
(1−𝑝)2
𝑝2
− 2πœƒ
1−𝑝
𝑝
(1−𝑝)/𝑝
1
1−𝑝
can compute π‘…πœƒΜ‚ (πœƒ) = 𝐸 [
𝑑 2 −2πœƒπ‘‘+πœƒ2
πœƒ2 +πœƒ
. Therefore, for the
1
] = πœƒ2 +πœƒ [𝐸(𝑋̅ 2 ) − 2πœƒπΈ(𝑋̅) + πœƒ 2 ] =
πœƒ2
1
+ πœƒ 2 ] = πœƒ2 +πœƒ [𝑛𝑝 πœƒ 2 + πœƒ 2 − 2πœƒ 2 + πœƒ 2 ] = (πœƒ2 +πœƒ)𝑛𝑝 =
= [(1−𝑝)/𝑝+1]𝑛𝑝 = [1/𝑝]𝑛𝑝2 =
(
πœƒ 2 +πœƒ
=
1−𝑝
𝑛𝑝
πœƒ
𝑛
= 𝑛. Similarly, for the estimator πœƒΜ‚ = 𝑛+1 𝑋̅ we
2
𝑛
𝑛
𝑋̅ ) −2πœƒ(
𝑋̅ )+πœƒ2
𝑛+1
𝑛+1
πœƒ2 +πœƒ
𝑛+πœƒ
] = β‹― = (πœƒ+1)(𝑛+1)2 .
Question #36: Let 𝑋1 , … , 𝑋𝑛 be a random sample of size 𝑛 from a Normal distribution such
that each 𝑋𝑖 ~𝑁(0, πœƒ). Find the asymptotic distribution of the MLE of the parameter πœƒ.
ο‚·
1
From the previous assignment, we know that πœƒΜ‚π‘€πΏπΈ = 𝑛 ∑𝑛𝑖=1 𝑋𝑖2 . We then use
Definition 9.4.5, which states that for large values of 𝑛, the MLE estimator is
distributed normal with mean πœƒ and variance 𝐢𝑅𝐿𝐡. That is, we have that
πœƒΜ‚π‘€πΏπΈ ~𝑁(πœƒ, 𝐢𝑅𝐿𝐡). This means that we must compute the Cramer-Rao Lower Bound.
Since 𝜏(πœƒ) = πœƒ, then 𝜏 ′ (πœƒ) = 1 and [𝜏 ′ (πœƒ)]2 = 1. Next, since we previously found that
𝑓(𝑋; πœƒ) =
1
1
√2πœ‹πœƒ
2
1
1
𝑒 −2πœƒπ‘‹ , then we have ln 𝑓(𝑋; πœƒ) = − 2 ln(2πœ‹πœƒ) − 2πœƒ 𝑋 2 so that
πœ•
𝑋2
1
πœ•2
2𝑋 2
1
1
𝑋2
ln 𝑓(𝑋; πœƒ) = − 2πœƒ + 2πœƒ2. We then find πœ•πœƒ2 ln 𝑓(𝑋; πœƒ) = 2πœƒ2 − 2πœƒ3 = 2πœƒ2 − πœƒ3 and take
πœ•πœƒ
1
𝑋2
1
1
the negative of its expected value to obtain −𝐸 [2πœƒ2 − πœƒ3 ] = − [2πœƒ2 − πœƒ3 𝐸(𝑋 2 )] =
1
1
πœƒ
2πœƒ2
(πœƒ + 02 ) −
3
1
1
1
= πœƒ2 − 2πœƒ2 = 2πœƒ2. This implies that 𝐢𝑅𝐿𝐡 =
2πœƒ2
𝑛
. Combining these
2
2πœƒ
facts reveals that the asymptotic distribution of the MLE is πœƒΜ‚π‘€πΏπΈ ~𝑁 (πœƒ, 𝑛 ). We can
transform this to get a standard normal distribution by noting that the random
variable
Μ‚ 𝑀𝐿𝐸 −πœƒ
πœƒ
πœƒ√2/𝑛
~𝑁(0,1) or large values of 𝑛. We could further reduce this by
multiplying through by the constant πœƒ so that
Μ‚ 𝑀𝐿𝐸 −πœƒ
πœƒ
√2/𝑛
~𝑁(0, πœƒ 2 ).
Chapter #10 – Sufficiency and Completeness
Question #6: Let 𝑋1 , … , 𝑋𝑛 be independent and each 𝑋𝑖 ~𝐡𝐼𝑁(π‘šπ‘– , 𝑝). Use the Factorization
Criterion to show that 𝑆 = ∑𝑛𝑖=1 𝑋𝑖 is sufficient for the unknown parameter 𝑝.
ο‚·
Since each 𝑋𝑖 ~𝐡𝐼𝑁(π‘šπ‘– , 𝑝), we know that the probability mass function is given by
𝑓(π‘₯; π‘šπ‘– , 𝑝) = (π‘šπ‘₯𝑖 )𝑝 π‘₯ (1 − 𝑝)π‘šπ‘– −π‘₯ 1{π‘₯ = 0, … , π‘šπ‘– }. We can then construct their joint
probability mass function due to the fact that they are independent as
𝑖
𝑓(π‘₯1 , … , π‘₯𝑛 ; π‘šπ‘– , 𝑝) = ∏𝑛𝑖=1 𝑓(π‘₯𝑖 ; π‘šπ‘– , 𝑝) = ∏𝑛𝑖=1 (π‘š
) 𝑝 π‘₯𝑖 (1 − 𝑝)π‘šπ‘– −π‘₯𝑖 1{π‘₯𝑖 = 0, … , π‘šπ‘– } =
π‘₯
𝑖
𝑛
𝑛
𝑛
𝑖
𝑖
[∏𝑛𝑖=1 (π‘š
)] 𝑝∑𝑖=1 π‘₯𝑖 (1 − 𝑝)∑𝑖=1 π‘šπ‘– −∑𝑖=1 π‘₯𝑖 1{π‘₯𝑖 = 0, … , π‘šπ‘– }. If 𝐢 = [∏𝑛𝑖=1 (π‘š
)], then we
π‘₯
π‘₯
𝑖
𝑖
𝑛
𝑛
𝑛
have that 𝐢𝑝∑𝑖=1 π‘₯𝑖 π‘ž ∑𝑖=1 π‘šπ‘– −∑𝑖=1 π‘₯𝑖 1{π‘₯𝑖 = 0, … , π‘šπ‘– } = 𝐢
𝑛
𝑝 ∑𝑖=1 π‘₯𝑖
𝐢 (π‘ž )
𝑛
𝑛
𝑝∑𝑖=1 π‘₯𝑖 π‘ž ∑𝑖=1 π‘šπ‘–
∑𝑛
π‘ž 𝑖=1 π‘₯𝑖
1{π‘₯𝑖 = 0, … , π‘šπ‘– } =
𝑛
π‘ž ∑𝑖=1 π‘šπ‘– 1{π‘₯𝑖 = 0, … , π‘šπ‘– }. But then if we define 𝑠 = ∑𝑛𝑖=1 π‘₯𝑖 , we have that
𝑝 𝑠
𝑛
𝑓(π‘₯1 , … , π‘₯𝑛 ; π‘šπ‘– , 𝑝) = 𝐢 (π‘ž ) π‘ž ∑𝑖=1 π‘šπ‘– 1{π‘₯𝑖 = 0, … , π‘šπ‘– } = 𝑔(𝑠; π‘šπ‘– , 𝑝)β„Ž(π‘₯1 , … , π‘₯𝑛 ).
𝑝 𝑠
Since
𝑛
𝑔(𝑠; π‘šπ‘– , 𝑝) = (π‘ž ) π‘ž ∑𝑖=1 π‘šπ‘– does not depend on π‘₯1 , … , π‘₯𝑛 except through 𝑠 = ∑𝑛𝑖=1 π‘₯𝑖
and β„Ž(π‘₯1 , … , π‘₯𝑛 ) = 𝐢1{π‘₯𝑖 = 0, … , π‘šπ‘– } does not involve 𝑝, the Factorization Criterion
guarantees that 𝑆 = ∑𝑛𝑖=1 𝑋𝑖 is sufficient for the unknown parameter 𝑝.
Question #7: Let 𝑋1 , … , 𝑋𝑛 be independent and each 𝑋𝑖 ~𝑁𝐡(π‘Ÿπ‘– , 𝑝). This means that each 𝑋𝑖
has probability mass function 𝑃(𝑋𝑖 = π‘₯) = (π‘Ÿπ‘₯−1
) π‘π‘Ÿπ‘– (1 − 𝑝)π‘₯−π‘Ÿπ‘– for π‘₯ = π‘Ÿπ‘– , π‘Ÿπ‘– + 1, π‘Ÿπ‘– + 2, …
−1
𝑖
Find a sufficient statistic for the unknown parameter 𝑝 using the Factorization Criterion.
ο‚·
As in the question above, we have that 𝑓(π‘₯1 , … , π‘₯𝑛 ; π‘Ÿπ‘– , 𝑝) = ∏𝑛𝑖=1 𝑓(π‘₯𝑖 ; π‘Ÿπ‘– , 𝑝) =
−1
∏𝑛𝑖=1 (π‘₯π‘Ÿπ‘–−1
) π‘π‘Ÿπ‘– (1 − 𝑝)π‘₯𝑖 −π‘Ÿπ‘– 1{π‘₯𝑖 = π‘Ÿπ‘– , π‘Ÿπ‘– + 1, … }. After applying the product operator,
𝑖
𝑛
𝑛
𝑛
−1
this becomes [∏𝑛𝑖=1 (π‘₯π‘Ÿπ‘–−1
)] 𝑝∑𝑖=1 π‘Ÿπ‘– π‘ž ∑𝑖=1 π‘₯𝑖 −∑𝑖=1 π‘Ÿπ‘– 1{π‘₯𝑖 = π‘Ÿπ‘– , π‘Ÿπ‘– + 1, … }. Then if we define
𝑖
𝐢=
−1
[∏𝑛𝑖=1 (π‘₯π‘Ÿπ‘–−1
)],
𝑖
𝑛
this expression becomes 𝐢
𝑛
𝑝∑𝑖=1 π‘Ÿπ‘– π‘ž ∑𝑖=1 π‘₯𝑖
𝑛
∑
π‘ž 𝑖=1 π‘Ÿπ‘–
1{π‘₯𝑖 = π‘Ÿπ‘– , π‘Ÿπ‘– + 1, … } =
𝑛
𝑝 ∑𝑖=1 π‘Ÿπ‘–
𝐢 (π‘ž )
joint
𝑛
π‘ž ∑𝑖=1 π‘₯𝑖 1{π‘₯𝑖 = π‘Ÿπ‘– , π‘Ÿπ‘– + 1, … }. Finally, if we let 𝑠 = ∑𝑛𝑖=1 π‘₯𝑖 , we have that the
mass
function
𝑛
𝑝 ∑𝑖=1 π‘Ÿπ‘–
is
𝑓(π‘₯1 , … , π‘₯𝑛 ; π‘Ÿπ‘– , 𝑝) = 𝐢 (π‘ž )
𝑛
𝑝 ∑𝑖=1 π‘Ÿπ‘–
𝑔(𝑠; π‘Ÿπ‘– , 𝑝)β„Ž(π‘₯1 , … , π‘₯𝑛 ). Since 𝑔(𝑠; π‘Ÿπ‘– , 𝑝) = (π‘ž )
π‘ž 𝑠 1{π‘₯𝑖 = π‘Ÿπ‘– , π‘Ÿπ‘– + 1, … } =
π‘ž 𝑠 does not depend on π‘₯1 , … , π‘₯𝑛
except through 𝑠 = ∑𝑛𝑖=1 π‘₯𝑖 and β„Ž(π‘₯1 , … , π‘₯𝑛 ) = 𝐢1{π‘₯𝑖 = π‘Ÿπ‘– , π‘Ÿπ‘– + 1, … } does not involve
𝑝, the Factorization Criterion guarantees that 𝑆 = ∑𝑛𝑖=1 𝑋𝑖 is sufficient for 𝑝.
Question #16: Let 𝑋1 , … , 𝑋𝑛 be independent and each 𝑋𝑖 ~𝑁𝐡(π‘Ÿπ‘– , 𝑝). This means that each 𝑋𝑖
has mass function 𝑃(𝑋𝑖 = π‘₯) = (π‘Ÿπ‘₯−1
) π‘π‘Ÿπ‘– (1 − 𝑝)π‘₯−π‘Ÿπ‘– for π‘₯ = π‘Ÿπ‘– , π‘Ÿπ‘– + 1, π‘Ÿπ‘– + 2, … Find the
−1
𝑖
Maximum Likelihood Estimator (MLE) of 𝑝 by maximizing the MLE of the sufficient statistic.
ο‚·
In
the
previous
question,
𝑛
𝑛
we
found
that
𝐿(𝑝) = 𝑓(π‘₯1 , … , π‘₯𝑛 ; π‘Ÿπ‘– , 𝑝) =
𝑛
−1
[∏𝑛𝑖=1 (π‘₯π‘Ÿπ‘–−1
)] 𝑝∑𝑖=1 π‘Ÿπ‘– (1 − 𝑝)∑𝑖=1 π‘₯𝑖 (1 − 𝑝)− ∑𝑖=1 π‘Ÿπ‘– . Taking the natural logarithm gives
𝑖
−1
ln[𝐿(𝑝)] = ∑𝑛𝑖=1 (π‘₯π‘Ÿπ‘–−1
) + ∑𝑛𝑖=1 π‘Ÿπ‘– ln(𝑝) + ∑𝑛𝑖=1 π‘₯𝑖 ln(1 − 𝑝) − ∑𝑛𝑖=1 π‘Ÿπ‘– ln⁑(1 − 𝑝).
𝑖
Then
differentiating the log likelihood function and equating to zero implies that
πœ•
ln[𝐿(𝑝)] =
πœ•π‘
∑𝑛
𝑖=1 π‘Ÿπ‘–
𝑝
−
∑𝑛
𝑖=1 π‘₯𝑖
1−𝑝
+
∑𝑛
𝑖=1 π‘Ÿπ‘–
1−𝑝
= 0 → (1 − 𝑝) ∑𝑛𝑖=1 π‘Ÿπ‘– − 𝑝 ∑𝑛𝑖=1 π‘₯𝑖 + 𝑝 ∑𝑛𝑖=1 π‘Ÿπ‘– = 0.
Then we have ∑𝑛𝑖=1 π‘Ÿπ‘– − 𝑝 ∑𝑛𝑖=1 π‘Ÿπ‘– − 𝑝 ∑𝑛𝑖=1 π‘₯𝑖 + 𝑝 ∑𝑛𝑖=1 π‘Ÿπ‘– = 0 → ∑𝑛𝑖=1 π‘Ÿπ‘– − 𝑝 ∑𝑛𝑖=1 π‘₯𝑖 = 0.
∑𝑛
π‘Ÿ
This implies that the Maximum Likelihood Estimator of 𝑝 is 𝑝̂ 𝑀𝐿𝐸 = ∑𝑛𝑖=1 π‘₯𝑖 .
𝑖=1 𝑖
Question #12: Let 𝑋1 , … , 𝑋𝑛 be independent and identically distributed from a two
parameter exponential distribution 𝐸𝑋𝑃(πœƒ, πœ‚) such that the probability density function is
1
𝑓(π‘₯; πœƒ, πœ‚) = πœƒ 𝑒
ο‚·
−π‘₯+πœ‚
πœƒ
1{π‘₯ > πœ‚}. Find jointly sufficient statistics for the parameters πœƒ and πœ‚.
Since the random variables are iid, their joint probability density function is thus
given by 𝑓(π‘₯1 , … , π‘₯𝑛 ; πœƒ, πœ‚) = ∏𝑛𝑖=1 𝑓(π‘₯𝑖 ; πœƒ, πœ‚) 1{π‘₯𝑖 > πœ‚} = ∏𝑛𝑖=1 πœƒ −1 𝑒
1
−π‘₯𝑖 +πœ‚
πœƒ
1{π‘₯𝑖 > πœ‚} =
𝑛
πœƒ −𝑛 𝑒 πœƒ(π‘›πœ‚−∑𝑖=1 π‘₯𝑖 ) [∏𝑛𝑖=1 1{π‘₯𝑖 > πœ‚}]. This then shows that 𝑆1 = ∑𝑛𝑖=1 𝑋𝑖 and 𝑆2 = 𝑋1:𝑛 are
jointly sufficient for πœƒ and πœ‚ by the Factorization Criterion with β„Ž(π‘₯1 , … , π‘₯𝑛 ) = 1 being
independent of the unknown parameters πœƒ and πœ‚ and 𝑔(𝑠1 , 𝑠2 ; πœƒ, πœ‚) =
1
πœƒ −𝑛 𝑒 πœƒ
(π‘›πœ‚−𝑠1 )
∏𝑛𝑖=1 1{𝑠2 > πœ‚} depending on π‘₯1 , … , π‘₯𝑛 only through 𝑆1 and 𝑆2 .
Question #13: Let 𝑋1 , … , 𝑋𝑛 be independent and identically distributed from a beta
distribution 𝐡𝐸𝑇𝐴(πœƒ1 , πœƒ2 ) such that the probability density function of each of these random
Γ(πœƒ +πœƒ )
variables is given by 𝑓(π‘₯; πœƒ1 , πœƒ2 ) = Γ(πœƒ 1)Γ(πœƒ2 ) π‘₯ πœƒ1 −1 (1 − π‘₯)πœƒ2 −1 whenever 0 < π‘₯ < 1. Find
1
2
jointly sufficient statistics for the unknown parameters πœƒ1 and πœƒ2 .
ο‚·
Since the random variables are iid, their joint density is given by 𝑓(π‘₯1 , … , π‘₯𝑛 ; πœƒ1 , πœƒ2 ) =
Γ(πœƒ +πœƒ )
πœƒ −1
1
2
∏𝑛𝑖=1 𝑓(π‘₯𝑖 ; πœƒ1 , πœƒ2 ) 1{0 < π‘₯𝑖 < 1} = ∏𝑛𝑖=1
π‘₯ 1 (1 − π‘₯𝑖 )πœƒ2 −1 1{0 < π‘₯𝑖 < 1} =
Γ(πœƒ )Γ(πœƒ ) 𝑖
1
2
Γ(πœƒ +πœƒ ) 𝑛
[Γ(πœƒ 1)Γ(πœƒ2 )] [π‘₯1 … π‘₯𝑛 ]πœƒ1 −1 [(1 − π‘₯1 ) … (1 − π‘₯𝑛 )]πœƒ2 −1 ∏𝑛𝑖=1 1{0 < π‘₯𝑖 < 1}.
1
2
This
then
shows that 𝑆1 = ∏𝑛𝑖=1 𝑋𝑖 and 𝑆2 = ∏𝑛𝑖=1(1 − 𝑋𝑖 ) are jointly sufficient for πœƒ1 and πœƒ2 by
the
Factorization
independent
of
Criterion
the
with
unknown
β„Ž(π‘₯1 , … , π‘₯𝑛 ) = ∏𝑛𝑖=1 1{0 < π‘₯𝑖 < 1}
parameters
and
being
𝑔(𝑠1 , 𝑠2 ; πœƒ1 , πœƒ2 ) =
Γ(πœƒ +πœƒ ) 𝑛
[Γ(πœƒ 1)Γ(πœƒ2 )] [𝑠1 ]πœƒ1 −1 [𝑠2 ]πœƒ2−1 depending on the observations only through 𝑆1 and 𝑆2 .
1
2
Question #18: Let 𝑋~𝑁(0, πœƒ) for πœƒ > 0. a) Show that 𝑋 2 is complete and sufficient for the
unknown parameter πœƒ, and b) show that 𝑁(0, πœƒ) is not a complete family.
a) Since 𝑋~𝑁(0, πœƒ), we know that 𝑓𝑋 (π‘₯; πœƒ) =
1
√2πœ‹πœƒ
π‘₯2
𝑒 −2πœƒ for π‘₯ ∈ ℝ. Therefore, by the
Regular Exponential Class (REC) Theorem, 𝑋 2 is complete and sufficient for πœƒ.
b) Since 𝑋~𝑁(0, πœƒ), we know that 𝐸 (𝑋) = 0 for all πœƒ > 0. Therefore, completeness fails
because we have a nontrivial unbiased estimator of 𝐸(𝑋) = 0.
Chapter #10 – Sufficiency and Completeness
Question #21: If 𝑋1 , … , 𝑋𝑛 is a random sample from a Bernoulli distribution such that each
𝑋𝑖 ~𝐡𝐸𝑅𝑁(𝑝) ≡ 𝐡𝐼𝑁(1, 𝑝) where 𝑝 is the unknown parameter to be estimated, find the
UMVUE for a) 𝜏(𝑝) = π‘‰π‘Žπ‘Ÿ(𝑋) = 𝑝(1 − 𝑝), and b) 𝜏(𝑝) = 𝑝2.
a) We first verify that the Bernoulli distribution is a member of the Regular Exponential
Class (REC) by noting that its density can be written as 𝑓(π‘₯; 𝑝) = 𝑝 π‘₯ (1 − 𝑝)1−π‘₯ =
𝑝π‘₯
π‘₯
π‘₯
𝑝π‘₯
(1−𝑝)π‘₯−1
= (1−𝑝)π‘₯ (1 − 𝑝) = (1−𝑝) (1 − 𝑝) = exp {ln [(1−𝑝) (1 − 𝑝)]}. This equality
implies
𝑓(π‘₯; 𝑝) = exp {π‘₯ ln (1−𝑝) + ln(1 − 𝑝)} = exp {π‘₯ ln (1−𝑝)} exp{ln⁑(1 − 𝑝)} =
𝑝
𝑝
𝑝
(1 − 𝑝) exp {π‘₯ ln (
𝑝
1−𝑝
𝑝
)} = 𝑐(𝑝) exp{𝑑1 (π‘₯)π‘ž1 (𝑝)}, so the Bernoulli distribution is a
member of the REC by Definition 10.4.2. We then use Theorem 10.4.2, which
guarantees the existence of sufficient statistics for distribution from the REC, to
construct the sufficient statistic 𝑆1 = ∑𝑛𝑖=1 𝑑1 (𝑋𝑖 ) = ∑𝑛𝑖=1 𝑋𝑖 . Next, we appeal to the
Rao-Blackwell Theorem in justifying the use of 𝑆1 (or any one-to-one function of it) in
our search for a UMVUE for π‘‰π‘Žπ‘Ÿ(𝑋) = 𝑝(1 − 𝑝). Our initial guess for an estimator is
𝑇 = 𝑋̅(1 − 𝑋̅), so we first compute that 𝐸(𝑇) = 𝐸[𝑋̅(1 − 𝑋̅)] = 𝐸(𝑋̅) − 𝐸(𝑋̅ 2 ) =
1
1
1
𝐸 (𝑛 ∑𝑛𝑖=1 𝑋𝑖 ) − [π‘‰π‘Žπ‘Ÿ(𝑋̅) + 𝐸(𝑋̅)2 ] = 𝑛 ∑𝑛𝑖=1 𝐸(𝑋𝑖 ) − 𝑛2 ∑𝑛𝑖=1 π‘‰π‘Žπ‘Ÿ(𝑋𝑖 ) + 𝐸(𝑋̅)2.
1
1
1
2
we have calculated that 𝐸(𝑇) = 𝑛 𝑛𝑝 − 𝑛2 𝑛𝑝(1 − 𝑝) − (𝑛 𝑛𝑝) = 𝑝 −
1
𝑝(1−𝑝)
𝑛
Thus,
− 𝑝2 =
𝑛−1
𝑝(1 − 𝑝) (1 − 𝑛) = 𝑝(1 − 𝑝) (
𝑛
𝑛
𝑛
), which implies that 𝑇 ∗ = 𝑛−1 𝑇 = 𝑛−1 [𝑋̅(1 − 𝑋̅]
will have expected value equal to π‘‰π‘Žπ‘Ÿ(𝑋) = 𝑝(1 − 𝑝). The Lehman-Scheffe Theorem
finally guarantees that 𝑇 ∗ is a UMVUE for π‘‰π‘Žπ‘Ÿ(𝑋) = 𝑝(1 − 𝑝) since it states that any
unbiased estimator which is a function of complete sufficient statistics is a UMVUE.
b) We note that for the complete sufficient statistic 𝑆1 = ∑𝑛𝑖=1 𝑋𝑖 , we have 𝐸(𝑆1 ) = 𝑛𝑝
and π‘‰π‘Žπ‘Ÿ(𝑆12 ) = 𝑛𝑝(1 − 𝑝) since 𝑆1 ~𝐡𝐼𝑁(𝑛, 𝑝), which is true because it is the sum of 𝑛
independent Bernoulli random variables. This implies 𝐸(𝑆12 ) = π‘‰π‘Žπ‘Ÿ(𝑆12 ) + 𝐸(𝑆1 )2 =
𝑛𝑝(1 − 𝑝) + (𝑛𝑝)2 = 𝑛𝑝(1 − 𝑝) + 𝑛2 𝑝2 . By the Lehman-Scheffe Theorem, we know
that we must use some function of the complete sufficient statistic 𝑆1 to construct a
UMVUE for the unknown parameter 𝑝2 . We note that for 𝑇 = 𝑆1 − 𝑆12 , we have 𝐸(𝑇) =
𝐸(𝑆1 ) − 𝐸(𝑆12 ) = 𝑛𝑝 − 𝑛𝑝(1 − 𝑝) + 𝑛2 𝑝2 = 𝑛𝑝 − 𝑛𝑝 + 𝑛𝑝2 + 𝑛2 𝑝2 = 𝑝2 (𝑛 + 𝑛2 ).
1
This implies that the statistic 𝑇 ∗ = 𝑛+𝑛2 𝑇 =
𝑆1 −𝑆12
𝑛+𝑛2
=
𝑛
∑𝑛
𝑖=1 𝑋𝑖 +(∑𝑖=1 𝑋𝑖 )
2
𝑛+𝑛2
will have
expected value equal to 𝑝2 , so it is a UMVUE by the Lehman-Scheffe Theorem.
Question #23: If 𝑋1 , … , 𝑋𝑛 is a random sample from a Normal distribution such that each
𝑋𝑖 ~𝑁(πœ‡, 9) where πœ‡ is unknown, find the UMVUE for a) the 95th percentile, and b) 𝑃(𝑋1 ≤ 𝑐),
where 𝑐 is a known constant. Hint: find the conditional distribution of 𝑋1 given 𝑋̅ = π‘₯, and
apply the Rao-Blackwell Theorem with 𝑇 = 𝑒(𝑋1 ), where define 𝑒(π‘₯) = 1{π‘₯ ≤ 𝑐}.
a) The 95th percentile of a random variable 𝑋 from a 𝑁(πœ‡, 9) distribution is the value of
π‘˜ such that 𝑃(𝑋 ≤ π‘˜) = 0.95 → 𝑃 (
𝑋−πœ‡
3
≤
π‘˜−πœ‡
3
) = 0.95 → 𝑃 (𝑍 ≤
π‘˜−πœ‡
3
) = 0.95 where
𝑍~𝑁(0,1). From tabulations of the standard normal distribution function Φ(𝑧), we
know that 𝑃(𝑍 ≤ 1.645) = 0.95, so we equate
π‘˜−πœ‡
3
= 1.645 → π‘˜ = 4.935 + πœ‡ = 𝜏(πœ‡).
This is what we wish to find a UMVUE for, but since the expectation of a constant is
that constant itself, we simply need to find a UMVUE for πœ‡. We begin by verifying that
the Normal distribution is a member of the Regular Exponential Class (REC) by noting
that the density of 𝑋~𝑁(πœ‡, 9) can be written as 𝑓(π‘₯; πœ‡) =
1
√18πœ‹
π‘₯2
exp {− 18 +
πœ‡π‘₯
9
1
√18πœ‹
1
exp {− 18 (π‘₯ − πœ‡)2 } =
πœ‡2
− 18}, where we have that 𝑑1 (π‘₯) = π‘₯ 2 and 𝑑2 (π‘₯) = π‘₯. Thus, the
Normal distribution is a member of the REC by Definition 10.4.2. We then use
Theorem 10.4.2, which guarantees the existence of sufficient statistics for
distribution from the REC, to construct the sufficient statistics 𝑆1 = ∑𝑛𝑖=1 𝑑1 (𝑋𝑖 ) =
∑𝑛𝑖=1 𝑋𝑖2 and 𝑆2 = ∑𝑛𝑖=1 𝑑2 (𝑋𝑖 ) = ∑𝑛𝑖=1 𝑋𝑖 . Since the sample mean is an unbiased
𝑆
estimator for the population mean, we have that 𝐸(𝑇) = 𝐸(𝑋̅) = 𝐸 ( 𝑛2 ) = πœ‡. Thus, an
unbiased estimator for 𝜏(πœ‡) = π‘˜ = 4.935 + πœ‡ is given by 𝑇 ∗ = 4.935 + 𝑋̅, which is
also a UMVUE of 𝜏(πœ‡) by the Lehmann-Scheffe Theorem.
𝑋1 −πœ‡
b) Note that we are trying to estimate 𝑃(𝑋1 ≤ 𝑐) = 𝑃 (
3
≤
𝑐−πœ‡
3
) = Φ(
𝑐−πœ‡
3
) = 𝜏(πœ‡),
where Φ: ℝ → (0,1) is the cumulative distribution function of 𝑍~𝑁(0,1). Since 𝜏(πœ‡) is
a nonlinear function of πœ‡, we cannot simply insert 𝑋̅ to obtain a UMVUE. To find an
unbiased estimator, we note that 𝑒(𝑋1 ) = 1{𝑋1 ≤ 𝑐} is unbiased for 𝜏(πœ‡) since we
have 𝐸[𝑒(𝑋1 )] = 𝐸[1{𝑋1 ≤ 𝑐}] = 𝑃(𝑋1 ≤ 𝑐) = 𝜏(πœ‡). But since it is not a function of
the complete sufficient statistic 𝑆2 = ∑𝑛𝑖=1 𝑋𝑖 , this estimator cannot be a UMVUE.
However, the Rao-Blackwell Theorem states that 𝐸[𝑒(𝑋1 )|𝑆2 ] = 𝐸[1{𝑋1 ≤ 𝑐}|𝑆2 ] will
also be unbiased and will be a function of 𝑆2 = ∑𝑛𝑖=1 𝑋𝑖 . The Lehmann-Scheffe
Theorem then guarantees that 𝐸[1{𝑋1 ≤ 𝑐}|𝑆2 ] will be a UMVUE. In order to find this,
we must compute the conditional distribution of 𝑋1 given 𝑆2 . We know that the
random variable 𝑆2 = ∑𝑛𝑖=1 𝑋𝑖 ~𝑁(π‘›πœ‡, 9𝑛) and that 𝑋1 = π‘₯, 𝑆2 = 𝑠 is equivalent to
2
𝑋1 =
π‘₯, ∑𝑛𝑖=2 𝑋𝑖
= 𝑠 − π‘₯. This implies that 𝑓𝑋1 |𝑆2 (π‘₯|𝑠) = β‹― =
𝑠
where πœ‡ ′ = 𝑛 and (𝜎 ′ )2 =
9(𝑛−1)
𝑛
𝐸[1{𝑋1 ≤ 𝑐}|𝑆2 ] = 𝑃(𝐴 ≤ 𝑐) = Φ (
𝑐−𝑠/𝑛
3√(𝑛−1)/𝑛
,
√2πœ‹πœŽ′
𝑠 9(𝑛−1)
. Therefore, if we let 𝐴~𝑁 (𝑛 ,
2
exp{−(π‘₯−πœ‡ ′ ) /2(𝜎′ ) }⁑
𝑛
) we have that
), which is a UMVUE for Φ (
𝑐−πœ‡
3
) = 𝜏(πœ‡).
Question #25: If 𝑋1 , … , 𝑋𝑛 is a random sample from the probability density function
𝑓(π‘₯; πœƒ) = πœƒπ‘₯ πœƒ−1 1{0 < π‘₯ < 1} where πœƒ > 0 is the unknown parameter, find the UMVUE for
1
1
a) 𝜏(πœƒ) = πœƒ by using the fact that 𝐸[− ln(𝑋)] = πœƒ, and b) the unknown parameter πœƒ.
a) We first verify that the density is a member of the REC by nothing that it can be
written
as
𝑓(π‘₯; πœƒ) = πœƒπ‘₯ πœƒ−1 = exp{ln[πœƒπ‘₯ πœƒ−1 ]} = exp{ln(πœƒ) + (πœƒ − 1) ln(π‘₯)} =
exp{ln(πœƒ)} exp{(πœƒ − 1) ln(π‘₯)} = πœƒ exp{(πœƒ − 1) ln(π‘₯)}, where 𝑑1 (π‘₯) = ln(π‘₯). We then
use Theorem 10.4.2, which guarantees the existence of sufficient statistics for REC
distributions, to construct the sufficient statistic 𝑆1 = ∑𝑛𝑖=1 𝑑1 (𝑋𝑖 ) = ∑𝑛𝑖=1 ln⁑(𝑋𝑖 ). Next,
we appeal to the Rao-Blackwell Theorem in justifying the use of 𝑆1 (or any one-to-one
1
function of it) in our search for a UMVUE for πœƒ. From the hint provided, we initially
guess
that
𝑇=
−𝑆1
𝑛
=
∑𝑛
𝑖=1 −ln⁑(𝑋𝑖 )
𝑛
and
check
that
𝐸(𝑇) = 𝐸 [
∑𝑛
𝑖=1 − ln(𝑋𝑖 )
𝑛
]=
1
1
1
1
∑𝑛 𝐸[− ln(𝑋𝑖 )] = 𝑛 = . The Lehman-Scheffe Theorem finally guarantees that
𝑛 𝑖=1
𝑛 πœƒ
πœƒ
1
1
𝑇 = − 𝑛 ∑𝑛𝑖=1 ln⁑(𝑋𝑖 ) is a UMVUE for πœƒ since it states that any unbiased estimator which
is a function of complete sufficient statistics is a UMVUE.
b) Any UMVUE of the unknown parameter πœƒ must be a function of the complete and
sufficient statistic 𝑆1 = ∑𝑛𝑖=1 ln⁑(𝑋𝑖 ) by the Lehman-Scheffe Theorem. We begin by
𝑛
noting that 𝐸(𝑆1 ) = 𝐸[∑𝑛𝑖=1 ln(𝑋𝑖 )] = ∑𝑛𝑖=1 𝐸[ln(𝑋𝑖 )] = − ∑𝑛𝑖=1 𝐸[− ln(𝑋𝑖 )] = − πœƒ, so
1
1
𝑆1
𝐸(𝑆1 )
we would like to be able to compute 𝐸 ( ) ≠
1
. However, this involves finding
1
𝐸 [− ln(𝑋)] since we know that 𝐸[− ln(𝑋)] = πœƒ. We do this by finding the distribution
1
of π‘Œ = −ln⁑(𝑋) using the CDF technique, which shows that π‘Œ~𝐸𝑋𝑃 (πœƒ) with density
1
𝑓(𝑦; πœƒ) = πœƒπ‘’ −πœƒπ‘₯ 1{π‘₯ > 0}. This is equivalent to π‘Œ~𝐺𝐴𝑀𝑀𝐴 (πœƒ , 1), so by the Moment
1
Generating Function technique, we see that 𝑆1 = ∑𝑛𝑖=1 ln⁑(𝑋𝑖 ) ~𝐺𝐴𝑀𝑀𝐴 (πœƒ , 𝑛). We can
1
πœƒπ‘›
∞1
thus calculate 𝐸 [− ln(𝑋)] = Γ(𝑛) ∫0
𝑛−1
𝑆
= ∑𝑛
𝑛−1
𝑖=1 ln⁑(𝑋𝑖 )
πœƒ
π‘₯ 𝑛−1 𝑒 −πœƒπ‘₯ 𝑑π‘₯ = β‹― = 𝑛−1, which implies that 𝑇 =
π‘₯
is an unbiased estimator of πœƒ. Then the Lehmann-Scheffe Theorem
guarantees that it is also a UMVUE for the unknown parameter πœƒ.
Question #31: If 𝑋1 , … , 𝑋𝑛 is a random sample from the probability density function
𝑓(π‘₯; πœƒ) = πœƒ(1 + π‘₯)−(1+πœƒ) 1{π‘₯ > 0} for unknown πœƒ > 0, find a) the MLE of πœƒ, b) a complete and
1
1
sufficient statistic for πœƒ, c) the CRLB for 𝜏(πœƒ) = πœƒ, d) the UMVUE for 𝜏(πœƒ) = πœƒ, e) the mean
and variance of the asymptotic normal distribution of the MLE, and f) the UMVUE for πœƒ.
a) We have 𝐿(πœƒ) = ∏𝑛𝑖=1 𝑓(π‘₯𝑖 ; πœƒ) = ∏𝑛𝑖=1 πœƒ(1 + π‘₯𝑖 )−(1+πœƒ) = πœƒ 𝑛 [∏𝑛𝑖=1(1 + π‘₯𝑖 )]−(1+πœƒ) so
that ln[𝐿(πœƒ)] = 𝑛 ln(πœƒ) − (1 + πœƒ) ∑𝑛𝑖=1 ln(1 + π‘₯𝑖 ). Then we have that
𝑛
πœƒ
− ∑𝑛𝑖=1 ln(1 + π‘₯𝑖 ) = 0 → πœƒ = ∑𝑛
𝑛
𝑖=1 ln(1+π‘₯𝑖 )
so that πœƒΜ‚π‘€πΏπΈ = ∑𝑛
𝑛
.
𝑖=1 ln(1+𝑋𝑖 )
πœ•
πœ•πœƒ
ln[𝐿(πœƒ)] =
b) To check that it is a member of the REC, we verify that we can write the probability
density function of 𝑋 as 𝑓(π‘₯; πœƒ) = πœƒ(1 + π‘₯)−(1+πœƒ) = exp{ln[πœƒ(1 + π‘₯)−(1+πœƒ) ]} =
exp{ln(πœƒ) − (1 + πœƒ) ln(1 + π‘₯)}, where 𝑑1 (π‘₯) = 1 and 𝑑2 (π‘₯) = ln⁑(1 + π‘₯). Thus, 𝑓(π‘₯; πœƒ)
is a member of the REC and 𝑆2 = ∑𝑛𝑖=1 𝑑2 (𝑋𝑖 ) = ∑𝑛𝑖=1 ln(1 + 𝑋𝑖 ) is a complete and
sufficient statistic for the unknown parameter πœƒ to be estimated.
1 2
1
1
c) Since 𝜏(πœƒ) = πœƒ, we have [𝜏 ′ (πœƒ)]2 = [− πœƒ2 ] = πœƒ4. Then we have that 𝑓(𝑋; πœƒ) =
πœ•
πœƒ(1 + 𝑋)−(1+πœƒ) so its log is ln 𝑓(𝑋; πœƒ) = ln(πœƒ) − (1 + πœƒ) ln(1 + 𝑋) and πœ•πœƒ ln 𝑓(𝑋; πœƒ) =
1
πœƒ
− ln(1 + 𝑋). Finally,
πœ•2
πœ•πœƒ2
ln 𝑓(𝑋; πœƒ) = −
1
πœƒ2
so that −𝐸 [
πœ•2
πœ•πœƒ2
(1/πœƒ4 )
ln 𝑓(𝑋; πœƒ)] =
1
πœƒ2
. These
1
results combined allow us to conclude that 𝐢𝑅𝐿𝐡 = 𝑛(1/πœƒ2 ) = π‘›πœƒ2 .
d) We previously verified that this density is a member of the REC and that the statistic
𝑆2 = ∑𝑛𝑖=1 ln(1 + 𝑋𝑖 ) is complete and sufficient for πœƒ. Next, we use the Rao-Blackwell
Theorem in justifying the use of 𝑆2 (or any one-to-one function of it) in our search for
1
a UMVUE for πœƒ. In order to compute 𝐸(𝑆2 ), we need to find the distribution of the
random variable π‘Œ = ln⁑(1 + 𝑋), which we do using the CDF technique. We thus have
that πΉπ‘Œ (𝑦) = 𝑃(π‘Œ ≤ 𝑦) = 𝑃(ln(1 + 𝑋) ≤ 𝑦) = 𝑃(𝑋 ≤ 𝑒 𝑦 − 1) = 𝐹𝑋 (𝑒 𝑦 − 1), so that
𝑑
𝑑
then π‘“π‘Œ (𝑦) = 𝑑𝑦 πΉπ‘Œ (𝑦) = 𝑑𝑦 𝐹𝑋 (𝑒 𝑦 − 1) = 𝑒 𝑦 𝑓𝑋 (𝑒 𝑦 − 1) = 𝑒 𝑦 [πœƒ(1 + 𝑒 𝑦 − 1)−(1+πœƒ) ] =
πœƒπ‘’ 𝑦 (𝑒 𝑦 )−(1+πœƒ) = πœƒπ‘’ 𝑦 𝑒 −(𝑦+πœƒπ‘¦) = πœƒπ‘’ 𝑦−𝑦−πœƒπ‘¦ = πœƒπ‘’ −πœƒπ‘¦
whenever
𝑦 > 0.
It
is
1
immediately clear that π‘Œ~𝐸𝑋𝑃(πœƒ), so that 𝐸(π‘Œ) = 𝐸[ln(1 + 𝑋)] = πœƒ. This allows us to
𝑛
find 𝐸(𝑆2 ) = 𝐸[∑𝑛𝑖=1 ln(1 + 𝑋𝑖 )] = ∑𝑛𝑖=1 𝐸[ln(1 + 𝑋𝑖 )] = πœƒ. Since we want an unbiased
1
estimator for πœƒ, it is clear that 𝑇 =
𝑆2
𝑛
1
= 𝑛 ∑𝑛𝑖=1 ln(1 + 𝑋𝑖 ) will suffice by the LST.
e) We previously found that the MLE for πœƒ is πœƒΜ‚π‘€πΏπΈ = ∑𝑛
𝑛
. From Chapter 9, we
𝑖=1 ln(1+𝑋𝑖 )
know that the MLE for some unknown parameter πœƒ has an asymptotic normal
distribution with πœ‡ = πœƒ and 𝜎 2 = 𝐢𝑅𝐿𝐡; that is, πœƒΜ‚π‘€πΏπΈ ~𝑁(πœƒ, 𝐢𝑅𝐿𝐡) for large 𝑛. We
must therefore find the Cramer-Rao Lower Bound, which can be easily done from the
work in part c) above with 𝜏(πœƒ) = πœƒ, so that 𝐢𝑅𝐿𝐡 =
πœƒ2
𝑛
. This means that we have
2
πœƒ
1
πœƒΜ‚π‘€πΏπΈ ~𝑁 (πœƒ, 𝑛 ) for large 𝑛. We can similarly argue for the MLE of 𝜏(πœƒ) = πœƒ, where we
1
1
see that πœƒΜƒπ‘€πΏπΈ = πœƒΜ‚
= 𝑛 ∑𝑛𝑖=1 ln(1 + 𝑋𝑖 ) by the Invariance Property of the Maximum
𝑀𝐿𝐸
Likelihood Estimator. Then using the work done in part c) above for the Cramer-Rao
1
1
Lower Bound, we can conclude that πœƒΜƒπ‘€πΏπΈ ~𝑁 (πœƒ , π‘›πœƒ2 ) for large 𝑛.
f) We previously verified that this density is a member of the REC and that the statistic
𝑛
𝑆2 = ∑𝑛𝑖=1 ln(1 + 𝑋𝑖 ) is complete and sufficient for πœƒ where 𝐸(𝑆2 ) = πœƒ . As in the
1
previous question, we have that 𝐸 (𝑆 ) = 𝐸 (∑𝑛
2
𝑇=
𝑛−1
𝑆2
= ∑𝑛
𝑛−1
𝑖=1 ln(1+𝑋𝑖 )
𝑖=1
1
πœƒ
) = 𝑛−1 which implies that
ln(1+𝑋 )
𝑖
is unbiased and a UMVUE for the unknown parameter πœƒ.
Chapter #11 – Interval Estimation
Question #5: If 𝑋1 , … , 𝑋𝑛 is a random sample from 𝑓𝑋 (π‘₯; πœ‚) = 𝑒 πœ‚−π‘₯ 1{π‘₯ > πœ‚} with πœ‚ unknown,
then a) show that 𝑄 = 𝑋1:𝑛 − πœ‚ is a pivotal quantity and find its distribution, and b) derive a
100𝛾% equal-tailed confidence interval for the unknown parameter πœ‚.
a) We first find the distribution of the smallest order statistic 𝑋1:𝑛 using the formula
𝑓1 (𝑦; πœ‚) = 𝑛𝑓𝑋 (𝑦; πœ‚)[1 − 𝐹𝑋 (𝑦; πœ‚)]𝑛−1 . We thus need the CDF of the population, which
is
given
π‘₯
π‘₯
π‘₯
𝐹𝑋 (π‘₯; πœ‚) = ∫πœ‚ 𝑓𝑋 (𝑑; πœ‚) 𝑑𝑑 = ∫πœ‚ 𝑒 πœ‚−𝑑 𝑑𝑑 = 𝑒 πœ‚ ∫πœ‚ 𝑒 −𝑑 𝑑𝑑 = 𝑒 πœ‚ [−𝑒 −𝑑 ]πœ‚π‘₯ =
by
𝑒 πœ‚ (−𝑒 −π‘₯ + 𝑒 −πœ‚ ) = 1 − 𝑒 πœ‚−π‘₯ whenever π‘₯ > πœ‚. We therefore have that 𝑓1 (𝑦; πœ‚) =
𝑛𝑒 πœ‚−𝑦 [1 − (1 − 𝑒 πœ‚−𝑦 )]𝑛−1 = 𝑛𝑒 πœ‚−𝑦 [𝑒 πœ‚−𝑦 ]𝑛−1 = 𝑛𝑒 πœ‚−𝑦 𝑒 π‘›πœ‚−πœ‚−𝑛𝑦+𝑦 = 𝑛𝑒 𝑛(πœ‚−𝑦) when
π‘₯ > πœ‚. Now that we have the density of 𝑋1:𝑛 , we can use the CDF technique to find the
density of 𝑄 = 𝑋1:𝑛 − πœ‚. Thus, we have 𝐹𝑄 (π‘ž) = 𝑃(𝑄 ≤ π‘ž) = 𝑃(𝑋1:𝑛 − πœ‚ ≤ π‘ž) =
𝑑
𝑃(𝑋1:𝑛 ≤ π‘ž + πœ‚) = 𝐹1 (π‘ž + πœ‚) so 𝑓𝑄 (π‘ž) = π‘‘π‘ž 𝐹1 (π‘ž + πœ‚) = 𝑓1 (π‘ž + πœ‚) = 𝑛𝑒 −π‘›π‘ž whenever
1
π‘ž + πœ‚ > πœ‚ → π‘ž > 0. This reveals that 𝑄 = 𝑋1:𝑛 − πœ‚~𝐸𝑋𝑃 (𝑛), so it is clearly a pivotal
quantity since it is a function of πœ‚ but its distribution does not depend on πœ‚.
b) We have 𝑃(π‘₯(1−𝛾)/2 < 𝑄 < π‘₯(1+𝛾)/2 ) = 𝛾 → 𝑃(π‘₯(1−𝛾)/2 < 𝑋1:𝑛 − πœ‚ < π‘₯(1+𝛾)/2 ) = 𝛾, so
after solving for the unknown parameter we obtain the 100𝛾% equal tailed
confidence interval 𝑃(𝑋1:𝑛 − π‘₯(1+𝛾)/2 < πœ‚ < 𝑋1:𝑛 − π‘₯(1−𝛾)/2 ) = 𝛾. This can also be
expressed as the random interval (𝑋1:𝑛 − π‘₯(1+𝛾)/2 , 𝑋1:𝑛 − π‘₯(1−𝛾)/2 ). Finally, we know
1
that the 𝐸𝑋𝑃 (𝑛) distribution has CDF 𝐹𝑄 (π‘ž) = 1 − 𝑒 −π‘›π‘ž so that 𝐹𝑄 (π‘₯𝛼 ) = 𝛼 implies
1
1 − 𝑒 −𝑛π‘₯𝛼 = 𝛼. We solve this last equality for π‘₯𝛼 = − 𝑛 ln⁑(1 − 𝛼). This means that the
1
confidence interval becomes (𝑋1:𝑛 + 𝑛 ln (
found by substituting 𝛼 =
1−𝛾
2
and 𝛼 =
1−𝛾
1+𝛾
2
2
1
1+𝛾
) , 𝑋1:𝑛 + 𝑛 ln (
2
)), where each term is
1
into the expression π‘₯𝛼 = − 𝑛 ln⁑(1 − 𝛼).
2
Question #7: If 𝑋1 , … , 𝑋𝑛 is a random sample from 𝑓𝑋 (π‘₯; πœƒ) = πœƒ2 π‘₯𝑒 −π‘₯
unknown parameter πœƒ, a) show that 𝑄 =
2
2 ∑𝑛
𝑖=1 𝑋𝑖
πœƒ2
~πœ’ 2 (2𝑛), b) use 𝑄 =
2 /πœƒ 2
1{π‘₯ > 0} with
2
2 ∑𝑛
𝑖=1 𝑋𝑖
πœƒ2
to derive an
equal-tailed 100𝛾% confidence interval for πœƒ, c) find a lower 100𝛾% confidence limit for
𝑃(𝑋 > 𝑑) = 𝑒 −𝑑
2 /πœƒ 2
, d) find an upper 100𝛾% confidence limit for the π‘π‘‘β„Ž percentile.
2
a) Since 𝑓𝑋 (π‘₯; πœƒ) = πœƒ2 π‘₯𝑒 −π‘₯
2 /πœƒ 2
1{π‘₯ > 0}, we know that 𝑋~π‘ŠπΈπΌ(πœƒ, 2). The CDF technique
then reveals that 𝑋 2 ~𝐸𝑋𝑃(πœƒ 2 ) so that ∑𝑛𝑖=1 𝑋𝑖2 ~𝐺𝐴𝑀𝑀𝐴(πœƒ 2 , 𝑛). A final application of
the CDF technique shows that
also shows that 𝑄 =
2
2 ∑𝑛
𝑖=1 𝑋𝑖
πœƒ2
2
πœƒ2
∑𝑛𝑖=1 𝑋𝑖2 ~πœ’ 2 (2𝑛), proving the desired result. This
is a pivotal quantity for the unknown parameter πœƒ.
2
2
b) We find that the confidence interval is 𝑃 (πœ’1−𝛾
(2𝑛) < 𝑄 < πœ’1+𝛾
(2𝑛)) = 𝛾 →
2
2
2 ∑𝑛 𝑋 2
2
2 ∑𝑛
𝑋2
2
2
𝑃 (πœ’1−𝛾
(2𝑛) < πœƒ2 ∑𝑛𝑖=1 𝑋𝑖2 < πœ’1+𝛾
(2𝑛)) = 𝛾 → 𝑃 (πœ’2 𝑖=1(2𝑛)𝑖 < πœƒ 2 < πœ’2 𝑖=1(2𝑛)𝑖 ) = 𝛾.
2
2
1+𝛾
2
1−𝛾
2
Taking square roots gives the desired random interval (√
2
2 ∑𝑛
𝑖=1 𝑋𝑖
2 (2𝑛)
πœ’1+𝛾
2
,√
2
2 ∑𝑛
𝑖=1 𝑋𝑖
2
2
2 ∑𝑛
𝑖=1 𝑋𝑖
c) From the work done above, a lower confidence limit for πœƒ is √
quantity 𝜏(πœƒ) = 𝑒 −𝑑
2
2 ∑𝑛
𝑖=1 𝑋𝑖
substitute √
πœ’π›Ύ2 (2𝑛)
2 /πœƒ 2
)⁑.
2 (2𝑛)
πœ’1−𝛾
πœ’π›Ύ2 (2𝑛)
. Since the
is a monotonically increasing function of πœƒ, we can simply
for πœƒ into the expression 𝜏(πœƒ) = 𝑒 −𝑑
2 /πœƒ 2
by Corollary 11.3.1.
d) We must solve the equation 𝑃(𝑋 > 𝑑𝑝 ) = 1 − 𝑝 for 𝑑𝑝 . From the question above, we
are given that 𝑃(𝑋 > 𝑑) = 𝑒 −𝑑
2 /πœƒ 2
2
2
so we must solve 𝑒 −𝑑𝑝 /πœƒ = 1 − 𝑝 for 𝑑𝑝 , which gives
2 ∑𝑛 𝑋 2
𝑑𝑝 = πœƒ√ln⁑(1 − 𝑝). By the same reasoning as above, we substitute √πœ’2 𝑖=1(2𝑛)𝑖 in for πœƒ
1−𝛾
2 ∑𝑛 𝑋 2
into the expression 𝑑𝑝 = 𝜏(πœƒ) = πœƒ√ln⁑(1 − 𝑝) to obtain 𝑑𝑝 = √πœ’2 𝑖=1(2𝑛)𝑖 ln⁑(1 − 𝑝).
1−𝛾
Question #8: If 𝑋1 , … , 𝑋𝑛 is a random sample from 𝑋~π‘ˆπ‘πΌπΉ(0, πœƒ) with πœƒ > 0 unknown and
𝑋𝑛:𝑛 is the largest order statistic, then a) find the probability that the random interval given
by (𝑋𝑛:𝑛 , 2𝑋𝑛:𝑛 ) contains πœƒ, and b) find the value of the constant 𝑐 such that the random
interval (𝑋𝑛:𝑛 , 𝑐𝑋𝑛:𝑛 ) is a 100(1 − 𝛼)% confidence interval for the parameter πœƒ.
a) We have that πœƒ ∈ (𝑋𝑛:𝑛 , 2𝑋𝑛:𝑛 ) if and only if πœƒ < 2𝑋𝑛:𝑛 , since the inequality πœƒ < 𝑋𝑛:𝑛
will always be true by the definition of the density. We must therefore compute
πœƒ
πœƒ
πœƒ
𝑛
𝑃(2𝑋𝑛:𝑛 > πœƒ) = 𝑃 (𝑋𝑛:𝑛 > 2) = 1 − 𝑃 (𝑋𝑛:𝑛 ≤ 2) = 1 − [𝑃 (𝑋𝑖 ≤ 2)] = 1 − 2−𝑛 .
b) As above, we have that 𝑃[πœƒ ∈ (𝑋𝑛:𝑛 , 𝑐𝑋𝑛:𝑛 )] = 1 − 𝑐 −𝑛 , so if we set this equal to 1 − 𝛼
and solve for the value of the constant, we obtain 1 − 𝑐 −𝑛 = 1 − 𝛼 → 𝑐 = 𝛼 −1/𝑛 .
Question #13: Let 𝑋1 , … , 𝑋𝑛 be a random sample from 𝑋~𝐺𝐴𝑀𝑀𝐴(πœƒ, πœ…) such that their
1
common distribution is 𝑓(π‘₯; πœƒ, πœ…) = πœƒπœ…Γ(πœ…) π‘₯ πœ…−1 𝑒 −π‘₯/πœƒ 1{π‘₯ > 0} with the parameter πœ… known
but πœƒ unknown. Derive a 100(1 − 𝛼)% equail-tailed confidence interval for πœƒ based on the
sufficient statistic for the unknown parameter πœƒ.
ο‚·
We begin by noting that the given density is a member of the Regular Exponential
Class (REC) since 𝑓(π‘₯; πœƒ, πœ…) = πœƒ −πœ… Γ(πœ…)−1 π‘₯ πœ…−1 𝑒 −π‘₯/πœƒ = 𝑐(πœƒ)β„Ž(π‘₯) exp{π‘ž1 (πœƒ)𝑑1 (π‘₯)}
where 𝑑1 (π‘₯) = π‘₯. Then we know that 𝑆 = ∑𝑛𝑖=1 𝑑1 (𝑋𝑖 ) = ∑𝑛𝑖=1 𝑋𝑖 is complete sufficient
for the unknown parameter πœƒ. Next, we need to create a pivotal quantity from 𝑆; from
2
2
the distribution in question 7 which is similar, we guess that 𝑄 = πœƒ 𝑆 = πœƒ ∑𝑛𝑖=1 𝑋𝑖
might be appropriate. We now derive the distribution of 𝑄 and, by showing that it is
simultaneously a function of πœƒ but its density does not depend on πœƒ, will verify that it
is a pivotal quantity. Since the 𝑋𝑖 ~𝐺𝐴𝑀𝑀𝐴(πœƒ, πœ…), we know that the random variable
𝐴 = ∑𝑛𝑖=1 𝑋𝑖 ~𝐺𝐴𝑀𝑀𝐴(πœƒ, π‘›πœ…).
𝑃 (∑𝑛𝑖=1 𝑋𝑖 ≤
1
π‘žπœƒ
2
) = 𝑃 (𝐴 ≤
π‘žπœƒ π‘›πœ…−1
( )
πœƒπ‘›πœ… Γ(π‘›πœ…) 2
π‘žπœƒ
πœƒ
π‘žπœƒ
2
Then
2
𝐹𝑄 (π‘ž) = 𝑃(𝑄 ≤ π‘ž) = 𝑃 (πœƒ ∑𝑛𝑖=1 𝑋𝑖 ≤ π‘ž) =
π‘žπœƒ
𝑑
1
which
π‘žπœƒ
π‘žπœƒ πœƒ
) = 𝐹𝐴 ( 2 ) so that 𝑓𝑄 (π‘ž) = π‘‘π‘ž 𝐹𝐴 ( 2 ) = 𝑓𝐴 ( 2 ) 2 =
𝑒 −( 2 )/πœƒ 2 = β‹― = 2π‘›πœ… Γ(π‘›πœ…) π‘ž π‘›πœ…−1 𝑒 −π‘ž/2 ,
shows
that
the
2
transformed random variable 𝑄 = πœƒ ∑𝑛𝑖=1 𝑋𝑖 ~𝐺𝐴𝑀𝑀𝐴(2, π‘›πœ…) ≡ πœ’ 2 (2π‘›πœ…). This allows
2
2
to compute 𝑃[πœ’π›Ό/2
(2π‘›πœ…) < 𝑄 < πœ’1−𝛼/2
(2π‘›πœ…)] = 1 − 𝛼, so after substituting in for 𝑄
2 ∑𝑛
𝑖=1 𝑋𝑖
and solving for πœƒ, we have 𝑃 [πœ’2
1−𝛼/2
2 ∑𝑛
𝑖=1 𝑋𝑖
< πœƒ < πœ’2
(2π‘›πœ…)
] = 1 − 𝛼, which is the desired
𝛼/2 (2π‘›πœ…)
100(1 − 𝛼)% equail-tailed confidence interval for πœƒ based on the sufficient statistic
𝑆 = ∑𝑛𝑖=1 𝑋𝑖 for the unknown parameter πœƒ.
Download