Stat 643 Problems, Fall 2000

advertisement
Stat 643 Problems, Fall 2000
Assignment 1
1. Let (H,T ) be a measurable space. Suppose that ., T and U are 5-finite positive
measures on this space with T ¥ U and U ¥ ..
.U
.T
a) Show that T ¥ . and .T
. . œ .U † . . a.s. ..
b) Show that if . ¥ U, then
..
.U
œ Š .U
.. ‹
•"
a.s. ..
2. (Cressie) Let H œ Ò • -ß -Ó for some - ž !, T be the set of Borel sets and T be
Lebesgue measure divided by 2-. For a subset of H, define the symmetric set for E as
• E œ Ö • =l= - E× and let V œ ÖE - T |E œ • E×Þ
a) Show that V is a sub 5-algebra of T .
b) Let \ be an integrable random variable. Find EÐ\lVÑÞ
3. Let H œ Ò!ß "Ó# , T be the set of Borel sets and T œ "# ? € "# . for ? a distribution
placing a unit point mass at the point Ð "# ß "Ñ and . 2-dimensional Lebesgue measure.
Consider the variable \Ð=Ñ œ =" and the sub 5-algebra of T generated by \ , V.
a) For E - T , find EÐME lVÑ œ T ÐElVÑ.
b) For ] Ð=Ñ œ =# , find EÐ] lVÑ.
4. Suppose that \ œ Ð\" ß \# ß ÞÞÞß \8 Ñ has independent components, where each \3 is
generated as follows. For independent random variables [3 µ normalÐ.ß "Ñ and
^3 µ Poisson Ð.Ñ, \3 œ [3 with probability : and \3 œ ^3 with probability " • :.
Suppose that . - Ò!ß_Ñ . Use the factorization theorem and find low dimensional
sufficient statistics in the cases that:
a) : is known to be "# , and
b) : - Ò!ß "Ó is unknown.
(In the first case the parameter space is @ œ Ö "# × ‚ Ò!ß_Ñ , while in the second it is
Ò!ß "Ó ‚ Ò!ß_Ñ.)
5. (Ferguson) Consider the probability distributions on Ðe" ß U" Ñ defined as follows. For
) œ Ð)ß :Ñ - @ œ e ‚ Ð!ß "Ñ and \ µ T) , suppose
T) Ò\ œ BÓ œ œ
Ð" • :Ñ:B•)
!
for B œ )ß ) € "ß ) € #ß ÞÞÞ
otherwise .
Let \" ß \# ß ÞÞÞß \8 be iid according to T) Þ
a) Argue that the family ÖT)8 ×)-@ is not dominated by a 5-finte measure, so that
the factorization theorem can not be applied to identify a sufficient statistic here.
b) Argue from first principles that the statistic X Ð\Ñ œ Ðmin\3 ß !\3 Ñ is
sufficient for the parameter ) .
1
c) Argue that the factorization theorem can be applied if ) is known (the first
factor of the parameter space @ is replaced by a single point) and identify a
sufficient statistic for this case.
d) Argue that if : is known, min\3 is a sufficient statistic.
6. Suppose that \ w is exponential with mean -•" (i.e. has density
0- ÐBÑ œ -expÐ • -BÑMÒB !Ó with respect to Lebesgue measure on e" , but that one
only observes \ œ \ w MÒ\ w ž "Ó. (There is interval censoring below B œ ".)
a) Consider the measure . on k œ Ö!× • Ð"ß_Ñ consisting of a point mass of 1
at 0 plus Lebesgue measure on (1,_). Give a formula for the R-N derivative of
T-\ wrt . on k .
b) Suppose that \" ß \# ß ÞÞÞß \8 are iid with the marginal distribution T-\ . Find a
2-dimensional sufficient statistic for this problem and argue that it is indeed
sufficient.
c) Argue carefully that your statistic from b) is minimal sufficient.
7. As a complement to Problem 6, consider this "random censoring" model. Suppose
that \ w is as in Problem 6. Suppose further that independent of \ w , ^ is exponential with
mean ". We will suppose that with [ œ mina\ w ß ^ b one observes
\ œ aMÒ[ œ \ w Óß [ b (one sees either \ w or a random censoring time less than \ w ). The
family of distributions of \ is absolutely continuous with respect to the product of
counting measure and Lebesgue measure on k œ Ö!ß "× ‚ e" (call this dominating
measure .).
a) Find an R-N derivative of T-\ wrt . on k .
b) Supose that \" ß \# ß ÞÞÞß \8 are iid with the marginal distribution T-\ . Find a
minimal sufficient statistic here. (Actually carefully proving minimal sufficiency
looks like it would be hard. Just make a good guess and say what needs to be
done to finish the proof.)
8. Schervish Exercise 2.8. And show that ^ is ancillary.
9. Schervish Exercise 2.13. (I think you need to assume that ÖB3 × spans e5 .)
10. Schervish Exercise 2.18.
11. Schervish Exercise 2.22. What is a first order ancillary statistic here?
12. Schervish Exercise 2.27.
2
Assignment 2
13. Schervish Exercise 2.34 (consider 8 iid observations).
14. Schervish Exercise 2.36.
15. Schervish Exercise 2.42.
16. Schervish Exercise 2.45.
17. Schervish Exercise 2.46.
18. Prove the following version of Theorem 24 (that Vardeman gave only a discrete case
argument for) À
Suppose that \ œ ÐX ß WÑ and the family c of distributions of \ on k œ g ‚ f is
dominated by the measure . œ / ‚ # (a product of 5-finite measures on g and f
)
respectively). With .T
. . Ð>ß =Ñ œ 0) Ð>ß =Ñ, let
1) Ð>Ñ œ ( 0) Ð>ß =Ñ.# Ð=Ñ and 0) Ð=l>Ñ œ
f
0) Ð>ß =Ñ
1) Ð>Ñ
Suppose that c is FI regular at )! and that 1w) Ð>Ñ œ 'f 0)w Ð>ß =Ñ. # Ð=Ñ a>. Then
M\ Ð)! Ñ
M X Ð )! Ñ
19. Let T! and T" be two distributions on k and 0! and 0" be densities of these with
respect to some dominating 5-finite measure .. Consider the parametric family of
distributions with parameter ) - Ò!ß "Ó and densities wrt . of the form
0) ÐBÑ œ Ð" • )Ñ0! ÐBÑ € )0" ÐBÑ ß
and suppose that \ has density 0) .
a) If KÐÖ!×Ñ œ KÐÖ"×Ñ œ "# , find the posterior distribution of )l\ .
b) Suppose now that K is the uniform distribution on Ò!ß "Ó. Find the posterior
distribution of )l\ .
i) What is the mean of this posterior distribution (one Bayesian way of
inventing a point estimator for :)?
ii) Consider the partition of @ into @! œ Ò!ß Þ&Ó and @" œ ÐÞ&ß "Þ!Ó. One
Bayesian way of inventing a test for H! :) - @! is to decide in favor of @! if
the posterior probability assigned to @! is at least .5. Describe as explicitly as
you can the subset of k where this is the case.
3
20. Consider the situation of problem 6 and maximum likelihood estimation of -.
a) Show that with Q œ #Ò\3 's equal to !Ó, in the event that Q œ 8 there is no
MLE of -, but that in all other cases there is a maximizer of the likelihood. Then
^ ,
argue that for any - ž !, with T- probability tending to 1, the MLE of -, say 8
exists.
b) Give a simple estimator of - based on Q alone. Prove that this estimator is
consistent for -. Then write down an explicit one-step Newton modification of
your estimator from a).
c) Discuss what numerical methods you could use to find the MLE from a) in the
event that it exists.
d) Give two forms of large sample confidence intervals for - based on the MLE
^ and two different approximations to M Ð-ÑÞ
8
"
21. Consider the situation of Problems 6 and 20. Below are some data artificially
generated from an exponential distribution.
.24, 3.20, .14, 1.86, .58, 1.15, .32, .66, 1.60, .34,
.61, .09, 1.18, 1.29, .23, .58, .11, 3.82, 2.53, .88
a) Plot the loglikelihood function for the uncensored data (the Bw values given
above). Give approximate 90% two-sided confidence intervals for - based on the
asymptotic ;# distribution for the LRT statistic for testing H! À - œ -! vs
H" À - Á -! and based on the asymptotic normal distribution of the MLE.
Now consider the censored data problem where any value less than 1 is reported as 0.
Modify the above data accordingly and do the following.
b) Plot the loglikelihood for the censored data (the derived B) values. How does
this function of - compare to the one from part a)? It might be informative to plot
these on the same set of axes.
c) It turns out (you might derive this fact) that
M" Ð- Ñ œ
expÐ • -Ñ
"
#
Œ
•ˆ" € - • expÐ • -щ Þ
#
- " • expÐ • -Ñ
Give two different approximate 90% confidence intervals for - based on the
asymptotic distribution of the MLE here. Then give an approximate 90% interval
based on inverting the LRTs of H! À - œ -! vs H" À - Á -! Þ
22. Suppose that \" ß \# ß \$ and \% are independent binomial random variables,
\3 µ binomial Ð8ß :3 Ñ. Consider the problem of testing H! À :" œ :# and :$ œ :% against
the alternative that H! does not hold.
a) Find the form of the LRT of these hypotheses and show that the log of the LRT
statistic is the sum of the logs of independent LRT statistics for H! À :" œ :# and
H! À :$ œ :% (a fact that might be useful in directly showing the ;# limit of the
LRT statistic under the null hypothesis).
b) Find the form of the Wald tests and show directly that the test statistic is
asymptotically ;# under the null hypothesis.
4
c) Find the form of the ;# (score) tests of these hypotheses and again show
directly that the test statistic is asymptotically ;# under the null hypothesis.
23. Suppose that \" ß \# ß ÞÞÞ are iid, each taking values in k œ Ö!ß "ß #× with R-N
derivative of T) wrt to counting measure on k
0) ÐBÑ œ
expÐB)Ñ
.
" € exp) € expÐ#)Ñ
a) Find an estimator of ) based on 8! œ ! MÒ\3 œ !Ó that is È8 consistent (i.e.
8
for which È8Ðs
) • )Ñ converges in distribution).
b) Find in more or less explicit form a "one step Newton modification " of your
estimator from a).
c) Prove directly that your estimator from b) is asymptotically normal with
variance "ÎM" Ð)Ñ. (With )^8 the estimator from a) and )˜ 8 the estimator from b),
3œ"
Pw8 Ðs
)8Ñ
˜)8 œ s
,
)8 •
s
ww
P8 Ð) 8 Ñ
‡
‡
and write Pw8 Ðs
) 8 Ñ œ Pw8 Ð)Ñ € Ðs
) 8 • )ÑPww8 Ð)Ñ € "# Ðs
) 8 • )Ñ# Pwww
8 Ð)8 Ñ for some )8
between s
) 8 and ).)
– - Ð!ß #Ñ the loglikelihood has a maximizer
d) Show that provided B
– • " € È • $B
– # € 'B
–€"
B
s
s
) 8 œ logŽ
• Þ
–
#Ð# • BÑ
s
– - Ð!ß #Ñ will be asymptotically
Prove that an estimator defined to be s
) 8 when B
normal with variance "ÎM" Ð)Ñ.
e) Show that the "observed information" and "expected information"
approximations lead to the same large sample confidence intervals for ). What do
s
these look like based on, say, s
)8?
By the way, a version of nearly everything in this problem works in any one parameter
exponential family.
24. Prove the discrete \ version of Theorem 30.
5
Assignment 3
25. (Optional) Consider the problem of Bayesian inference for the binomial parameter :.
In particular, for sake of convenience, consider the Uniform Ð!ß "Ñ (BetaÐ!ß " Ñ for
! œ " œ ") prior distribution.
(a) It is possible to argue from reasonably elementary principles that in this
binomial context, where @ œ Ð!ß "Ñ, the Beta posteriors have a consistency
property. That is, simple arguments can be used to show that for any fixed :! and
any % ž 0, for \8 µ binomial Ð8ß :! Ñ, the random variable
]8 œ (
:! €%
:! •%
"
:!€\8 •" Ð" • :Ñ" €Ð8•\8 Ñ•" .:
FÐ! € \8 ß " € Ð8 • \8 ÑÑ
(which is the posterior probability assigned to the interval Ð:! • %ß :! € % Ñ)
converges in :! probability to 1 as 8 p _. This problem is meant to lead you
through this argument. Let % ž 0 and $ ž 0.
i) Argue that there exists 7 such that if 8
7, ¹ B88 •
!€B8
!€" €8 ¹
•
%
3
aB8 œ !ß "ß ÞÞÞß 8Þ
8 ÑÐ" €8•B8 Ñ
w
ii) Note that the posterior variance is Ð!€Ð!"€B
€8Ñ# Ð!€" €8€"Ñ Þ Argue there is an 7
such that if 8 7w the probability that the posterior assigns to
% !€B8
%
8
Š !!€€B
" €8 • 3 ß !€" €8 € 3 ‹is at least " • $ aB8 œ !ß "ß ÞÞÞß 8.
iii) Argue that there is an 7ww such that if 8
¸ \88 • :! ¸ • $% is at least " • $ .
7ww the :! probability that
Then note that if 8 maxÐ7ß 7w ß 7ww Ñ i) and ii) together imply that the posterior
probability assigned to ˆ B88 • #$% ß B88 € #$% ‰is at least " • $ for any realization B8 .
Then provided ¸ B88 • :! ¸ • $% the posterior probability assigned to
Ð:! • %ß :! € %Ñ is also at least " • $ Þ But iii) says this happens with :!
probability at least " • $ . That is, for large 8, with :! probability at least " • $ ,
]8 " • $ Þ Since $ is arbitrary, (and ]8 Ÿ ") we have the convergence of ]8 to
1 in :! probability.
(b) Corollary 45 suggests that in an iid model, posterior densities for large 8 tend
to look normal (with means and variances related to the likelihood material). The
posteriors in this binomial problem are Beta Ð! € B8 ß " € Ð8 • B8 ÑÑ (and we can
think of \8 µ binomial Ð8ß :! Ñ as derived as the sum of 8 iid Bernoulli Ð:! Ñ
variables). So we ought to expect Beta distributions for large parameter values to
look roughly normal. To illustrate this (and thus essentially what Corollary 45
says in this case), do the following. For 3 œ Þ$ (for example ... any other value
would do as well), consider the Beta Ð! € 83ß " € 8Ð" • 3ÑÑ (posterior)
distributions for 8 œ "!ß #!ß %! and "!!. For :8 µ Beta Ð! € 83ß " € 8Ð" • 3ÑÑ
plot the probability densities for the variables
6
8
Ê 3Ð" • 3Ñ a:8 • 3b
on a single set of axes along with the standard normal density. Note that if [ has
‰
pdf 0 Ð † Ñ, then +[ € , has pdf 1Ð † Ñ œ +" 0 ˆ †•,
+ . (Your plots are translated and
rescaled posterior densities of : based on possible observed values B8 œ Þ$8.)
If this is any help in doing this plotting, I tried to calculate values of the Beta
function using MathCad and got the following:
ÐFÐ%ß )ÑÑ•" œ "Þ$# ‚ "!$ ß ÐFÐ(ß "&ÑÑ•" œ )Þ"% ‚ "!& ß
ÐFÐ"$ß #*ÑÑ•" œ #Þ#*" ‚ "!"" and ÐFÐ$"ß ("ÑÑ•" œ #Þ*'( ‚ "!#( .
26. Problem 12, parts a)-c) Schervish, page 534.
27. Consider the following model. Given parameters -" ß ÞÞÞß -R variables \" ß ÞÞÞß Þ\R
are independent Poisson variables, \3 µ PoissonÐ-3 ÑÞ Q is a parameter taking values in
Ö"ß #ß ÞÞÞß R × and if 3 Ÿ Q ß -3 œ ." , while if 3 ž Q ß -3 œ .# Þ (Q is the number of
Poisson means that are the same as that of the first observation.) With parameter vector
) œ ÐQß ." ß .# Ñ belonging to @ œ Ö"ß #ß ÞÞÞß R × ‚ Ð!ß_Ñ‚Ð!ß_Ñ we wish to make
inference about Q based on \" ß ÞÞÞß Þ\R in a Bayesian framework.
As matters of notational convenience, let W7 œ ! \3 and X œ W R .
7
3œ"
a) If, for example, a prior distribution K on @ is constructed by taking Q uniform
on Ö"ß #ß ÞÞÞß R × independent of ." exponential with mean 1, independent of .#
exponential with mean 1ß it is possible to explicitly find the (marginal) posterior
of Q given that \" œ B" ß ÞÞÞß \R œ BR . Don't actually bother to finish the
somewhat messy calculations needed to do this, but show that this is possible
(indicate clearly why appropriate integrals can be evaluated explicitly).
b) Suppose now that K is constructed by taking Q uniform on Ö"ß #ß ÞÞÞß R ×
independent of Ð." ß .# Ñ with joint density 1Ð † ß † Ñ wrt Lebesgue measure on
(0,_) ‚ (0,_). Describe in as much detail as possible a SSS based method for
approximating the posterior of Q , given \" œ B" ß ÞÞÞß \R œ BR . (Give the
necessary conditionals up to multiplicative constants, say how you're going to use
them and what you'll do with any vectors you produce by simulation.)
28. Consider the simple two-dimensional discrete distribution for ) œ Ð)" ß )# Ñ given in
the table below.
7
)#
)"
4
3
2
1
1
0
0
.2
.2
2
0
0
.1
.1
3
.2
.05
0
0
4
.1
.05
0
0
a) This is obviously a very simple setup where anything one might wish to know
or say about the distribution is easily derivable from the table. However, for sake
of example, suppose that one wished to use SSS to approximate
U œ T Ò)" Ÿ #Ó œ Þ'. Argue carefully that Gibbs Sampling (SSS) will fail here to
produce a correct estimate of U. What is it about the transition matrix that
describes SSS here that "causes" this failure?
b) Consider the very simple version of the Metropolis-Hastings algorithm where
the matrix X œ Ð>)w ß) Ñ is ") 1 (for 1 an ) ‚ ) matrix of "'s) and ) w and ) are vectors
for which the corresponding probabilities in the table above are positive. Find the
transition matrix T and verify that it has the distribution above as it invariant
distribution.
29. Suppose for sake of argument that I am interested in the properties of the discrete
distribution with probability mass function
0 ÐBÑ œ œ
|
5 |sinB
#B
!
for B œ "ß #ß á
otherwise
but, sadly, I don't know 5.
a) Describe a MCMC method I could use to approximate the mean of this
distribution.
(b) One doesn't really have to resort to MCMC here. One could generate iid
observations from 0 ÐBÑ (instead of a MC with stationary distribution 0 ÐBÑ) using
the "rejection method." (And then rely on the LLN to produce an approximate
mean for the distribution.) Prove that the following algorithm (based on iid
Uniform Ð!ß "Ñ random variables and samples from a geometric distribution) will
produce \ from the this distribution.
Step 1 Generate \ ‡ from the geometric distribution with probability mass
function
2ÐBÑ œ œ #B
!
"
for B œ "ß #ß á
otherwise
Step 2 Generate Y that is Uniform Ð!ß "Ñ
B|
Step 3 Note that 2ÐBÑ |sin
#B . If Y 2ÐBÑ •
return to Step 1.
|sinB|
#B
set \ œ \ ‡ . Otherwise
8
30. Suppose that f is a convex subset of Ò!ß_Ñ 5 . Argue that there exists a finite @
decision problem that has f as its set of risk vectors. (Consider problems where \ is
degenerate and carries no information.)
31. Consider the two state decision problem with @ œ Ö"ß #×, T" the Bernoulli ˆ "% ‰
distribution and T# the Bernoulli ˆ "# ‰distribution, T œ @ and PÐ)ß +Ñ œ MÒ) Á +ÓÞ
a) Find the set of risk vectors for the four nonrandomized decision rules. Plot
these in the plane. Sketch f , the risk set for this problem.
b) For this problem, show explicitly that any element of W‡ has a corresponding
element of W‡ with identical risk vector and vice versa.
c) Identify the set of all admissible risk vectors for this problem. Is there a
minimal complete class for this decision problem? If there is one, what is it?
d) For each : - Ò!ß "Ó, identify those risk vectors that are Bayes versus the prior
1 œ Ð:ß " • :Ñ. For which priors are there more than one Bayes rule?
e) Verify directly that the prescription "choose an action that minimizes the
posterior expected loss" produces a Bayes rule versus the prior 1 œ Ð "# ß "# ÑÞ
32. Consider a two state decision problem with @ œ Ö"ß #×, where the observable
\ œ Ð\" ß \# Ñ has iid Bernoulli ˆ "% ‰coordinates if ) œ " and iid Bernoulli ˆ "# ‰coordinates
if ) œ #. Suppose that T œ @ and PÐ)ß +Ñ œ MÒ) Á +Ó. Consider the behavioral decision
rule 9B defined by
9B ÐÖ"×Ñ œ "
if B" œ !, and
"
"
9B ÐÖ"×Ñ œ # and 9B ÐÖ#×Ñ œ #
if B" œ ".
a) Show that 9B is inadmissible by finding a rule with a better risk function. (It
may be helpful to figure out what the risk set is for this problem, in a manner
similar to what you did in problem 31.)
b) Use the construction from Result 61 and find a behavioral decision rule that is
a function of the sufficient statistic \" € \# and is risk equivalent to 9B Þ (Note
that this rule is inadmissible.)
33. Consider the squared error loss estimation of : - Ð!ß "Ñ, based on \ µ binomial
Ð8ß :Ñ, and the two nonrandomized decision rules $" ÐBÑ œ 8B and $# ÐBÑ œ "# ˆ 8B € #" ‰Þ Let
< be a randomized decision function that chooses $" with probability "# and $# with
probability "# Þ
a) Write out expressions for the risk functions of $" , $# and <.
b) Find a behavioral rule that is risk equivalent to <.
c) Identify a nonrandomized estimator that is strictly better than < or 9.
34. Suppose that @ œ @" ‚ @# and that a decision rule 9 is such that for each )# ß 9 is
admissible when the parameter space is @" ‚ Ö)# ×. Show that 9 is then admissible when
the parameter space is @.
9
35. Suppose that AÐ)Ñ ž ! a)Þ Show that 9 is admissible with loss function PÐ)ß +Ñ iff it
is admissible with loss function AÐ)ÑPÐ)ß +Ñ.
36. Problem 9, page 209 of Schervish.
10
Assignment 4
37. (Ferguson) Prove or give a counterexample: If V" and V# are complete classes of
decision rules, then V" • V# is essentially complete.
38. (Ferguson and others) Suppose that \ µ binomial Ð8ß :Ñ and one wishes to estimate
: - Ð!ß "Ñ. Suppose first that PÐ:ß +Ñ œ :•" Ð" • :Ñ•" Ð: • +Ñ# Þ
a) Show that \8 is Bayes versus the uniform prior on Ð!ß "Ñ.
b) Argue that \8 is admissible in this decision problem.
c) Show that \8 is minimax and identify a least favorable prior
Now consider ordinary squared error loss, PÐ:ß +Ñ œ Ð: • +Ñ# .
d) Apply the result of problem 35 and prove that \8 is admissible under this loss
function as well.
39. Consider a two state decision problem where @ œ T œ Ö!ß "×, T0 and T" have
respective densities with respect to a dominating 5-finite measure ., 0! and 0" and the
loss function is PÐ)ß +Ñ.
a) For K an arbitrary prior distribution, find a formal Bayes rule versus KÞ
b) Specialize your result from a) to the case where PÐ)ß +Ñ œ MÒ) Á +Ó. What
connection does the form of these rules have to the theory of simple versus simple
hypothesis testing?
40. Suppose that \ µ Bernoulli Ð:Ñ and that one wishes to estimate : with loss
PÐ:ß +Ñ œ l: • +l. Consider the estimator $ with $ Ð!Ñ œ "% and $ Ð"Ñ œ $% .
a) Write out the risk function for $ and show that VÐ:ß $ Ñ Ÿ "% Þ
b) Show that there is a prior distribution placing all its mass on Ö!ß "# ß "× against
which $ is Bayes.
c) Prove that $ is minimax in this problem and identify a least favorable prior.
41. Consider a decision problem where T) is the Normal Ð)ß "Ñ distribution on k =e" ,
T œ Ö!ß "× and PÐ)ß +Ñ œ MÒ+ œ !ÓMÒ) ž &Ó € MÒ+ œ "ÓMÒ) Ÿ &Ó.
a) If @ œ Ð • _ß &Ó • Ò'ß_Ñ guess what prior distribution is least favorable, find
the corresponding Bayes decision rule and prove that it is minimax.
b) If @ œ e" , guess what decision rule might be minimax, find its risk function
and prove that it is minimax.
42. (Ferguson and others) Consider (inverse mean) weighted squared error loss
estimation of -, the mean of a Poisson distribution. That is, let A œ Ð!ß_Ñß T œ Aß Tbe the Poisson distribution on k œ Ö!ß "ß #ß $ß ÞÞÞ× and PÐ-ß +Ñ œ -•" Ð- • +Ñ# Þ Let
$Ð\Ñ œ \Þ
a) Show that $ is an equalizer rule.
b) Show that $ is generalized Bayes versus Lebesgue measure on A.
c) Find the Bayes estimators wrt the >Ð!," Ñ priors on A.
d) Prove that $ is minimax for this problem.
11
43. Problem 14, page 340 Schervish.
44. Let R œ ÐR" ß R# ß ÞÞÞß R5 Ñ be multinomial Ð8ß :" ß :# ß ÞÞÞß :5 Ñ where !:3 œ "Þ
a) Find the form of the Fisher information matrix based on the parameter :.
b) Suppose that :3 Ð)Ñ for 3 œ "ß #ß ÞÞÞß 5 are differentiable functions of a real
parameter ) - @, and open interval, where each :3 Ð)Ñ ! and !:3 Ð)Ñ œ "Þ
Suppose that 2ÐC" ß C# ß ÞÞÞß C5 Ñ is a continuous real-valued function with continuous
first partial derivatives and define ;Ð)Ñ œ 2Ð:" Ð)Ñß :# Ð)Ñß ÞÞÞß :5 Ð)ÑÑ. Show that
the information bound for unbiased estimators of ;Ð)Ñ in this context is
5
Ð; w Ð)ÑÑ#
.
where M" Ð)Ñ œ ":3 Ð)ÑŒ log:3 Ð)Ñ• Þ
8M" Ð)Ñ
.)
3œ"
#
45. Show that the C-R inequality is not changed by a smooth reparameterization. That is,
suppose that c œ ÖT) × is dominated by a 5-finite measure . and satisfies
i) 0) ÐBÑ ž ! for all ) and B,
ii) for all B, ..) 0) ÐBÑ exists and is finite everywhere on @ § e" and
iii) for any statistic $ with E) l$ Ð\Ñl • _ for all ), E) $ Ð\Ñ can be differentiated
under the integral sign at all points of @.
Let 2 be a function from @ to e" such that 2w is continuous and nonvanishing on @. Let
( œ 2Ð)Ñ and define U( œ T) . Show that the information inequality bound obtained
from ÖU( × evaluated at ( œ 2Ð)Ñ is the same as the bound obtained from c .
46. As an application of Lemma 88 (or 88w ) consider the following. Suppose that
\ µ UÐ)" ß )# Ñ and consider unbiased estimators of # Ð)" ß )# Ñ œ )" €#)# . For )" • )"w • )#
and )" • )#w • )# , apply Lemma 88 with 1" ÐBÑ œ " •
0Ð)w ß)# Ñ ÐBÑ
"
0Ð)" ß)# Ñ ÐBÑ
and 1# ÐBÑ œ " •
0Ð)" ß)w Ñ ÐBÑ
0Ð)" ß)# Ñ ÐBÑ Þ
#
What does Lemma 88 then say about the variance of an unbiased estimator of # ()" ß )# Ñ?
(If you can do it, sup over values of )"w and )#w . I've not tried this, so am not sure how it
comes out.)
47. Problem 11, page 389 Schervish.
48. Consider the estimation problem with 8 iid T) observations, where T) is the
exponential distribution with mean ). Let Z be the group of scale transformations on
k œ Ð!ß_Ñ8 ß Z œ Ö1- l- ž !× where 1- ÐBÑ œ -BÞ
a) Show that the estimation problem with loss PÐ)ß +Ñ œ ÐlogÐ+Î)ÑÑ# is invariant
under Z and say what relationship any equivariant nonrandomized decision rule
must satisfy.
b) Show that the estimation problem with loss PÐ)ß +Ñ œ Ð) • +Ñ# Î) # is invariant
under Z , and say what relationship any equivariant nonrandomized estimator must
satisfy.
12
c) Find the generalized Bayes estimator of ) in situation b) if the "prior" has
density wrt Lebesgue measure 1Ð)Ñ œ )•" MÒ) ž !Ó. Argue that this estimator is
the best equivariant estimator in the situation of b).
13
Assignment 5
49. Problem 19, page 341 Schervish. Also, find the MRE estimator of ) under squared
error loss for this model.
50. Let 0 Ð?ß @Ñ be the bivariate probability density of the distribution uniform on the
square in Ð?ß @Ñ-space with corners at ÐÈ#ß !Ñß Ð!ß È#Ñß Ð • È#ß !Ñ and Ð!ß • È#Ñ.
Suppose that \ œ Ð\" ß \# Ñ has bivariate probability density
0 ÐBl)Ñ œ 0 ÐB" • )ß B# • ) Ñ. Find explicitly the best location equivariant estimator of )
under squared error loss. (It may well help you to visualize this problem to "sketch" the
joint density here for ) œ "(.)
51. Consider again the scenario of problem 31. There you sketched the risk set f . Here
sketch the corresponding set i .
52. Prove the following "filling-in" lemma:
Lemma Suppose that 1! and 1" are two distinct, positive probability densities defined on
an interval in e" Þ If the ratio 1" Î1! is nondecreasing in a real-valued function X ÐBÑ, then
the family of densities Ö1! | ! - Ò!ß "Ó× for 1! œ !1" € Ð" • !Ñ1# has the MLR property
in X ÐBÑ.
53. Consider the two distributions T! and T" on k œ Ð!ß "Ñ with densities wrt Lebesgue
measure
0! ÐBÑ œ " and 0" ÐBÑ œ $B# .
a) Find a most powerful level ! œ Þ# test of H! À \ µ T! vs H" À \ µ T" .
b) Plot both the !-" loss risk set f and the set i œ Ö"9 Ð)! Ñß " Ð)" Ñ× for the simple
versus testing problem involving 0! and 0" . What test is Bayes versus a uniform
prior? This test is best of what size?
c) Take the result of Problem 52 as given. Consider the family of mixture
distributions c œ ÖT) × where for ) - Ò!ß "Ó, T) œ Ð" • )ÑT! € ) T" . In this
family find a UMP level ! œ Þ# test of the hypothesis H! À ) Ÿ Þ& vs H" À ) ž Þ&
based on a single observation \ . Argue carefully that your test is really UMP.
54. Consider a composite versus composite testing problem in a family of distributions
c œ ÖT) × dominated by the 5-finite measure ., and specifically a Bayesian decisiontheoretic approach to this problem with prior distribution K under !-" loss.
a) Show that if KÐ@! Ñ œ ! then 9ÐBÑ » " is Bayes, while if KÐ@1 Ñ œ ! then
9ÐBÑ » 0 is Bayes.
b) Show that if KÐ@! Ñ ž ! and KÐ@1 Ñ ž ! then the Bayes test against K has
Neyman-Pearson form for densities 1! and 1" on k defined by
14
1! ÐBÑ œ
' 0) ÐBÑ.KÐ)Ñ
@!
KÐ@! Ñ
and 11 ÐBÑ œ
' 0) ÐBÑ.KÐ)Ñ
@1
KÐ@1 Ñ
.
c) Suppose that \ is Normal (),1). Find Bayes tests of H! À ) œ ! vs H" À ) Á !
i) supposing that K is the Normal Ð!ß 5# Ñ distribution, and
ii) supposing that K is "# ÐR € ?Ñ, for R the Normal Ð!ß 5# Ñ distribution,
and ? a point mass distribution at 0.
!
55. Fix ! - Ð!ß Þ&Ñ and - - Ð #•#
! ß !Ñ. Let @ œ Ö • "× • Ò!ß "Ó and consider the discrete
distributions with probability mass functions as below.
B
•#
•"
!
"
#
!
"
"
!
) œ •"
!
#
# •!
# •!
#
"•- ‰ˆ "
"•- ‰
"•- ‰ˆ "
‰
ˆ
ˆ
‰
a
) Á • " )- ˆ "•
•
!
!
•
!
"
•
)b!
#
"•!
"•!
#
Find the size ! likelihood ratio test of H! À ) œ • " vs H" À ) Á • ". Show that the test
9ÐBÑ œ MÒB œ !Ó is of size ! and is strictly more powerful that the LRT whatever be ).
(This simple example shows that LRTs need not necessarily be in any sense optimal.)
56. Suppose that \ is Poisson with mean ). Find the UMP test of size ! œ Þ!& of
H! À ) Ÿ % or ) 10 vs H" À % • ) • "!. The following table of Poisson probabilities
T) Ò\ œ BÓ will be useful in this enterprise.
B
)œ%
) œ "!
B
)œ%
) œ "!
!
Þ!")$
Þ!!!!
"$
Þ!!!#
Þ!(#*
"
Þ!($$
Þ!!!&
"%
Þ!!!"
Þ!&#"
#
Þ"%'&
Þ!!#$
"&
Þ!!!!
Þ!$%(
$
Þ"*&%
Þ!!('
"'
%
Þ"*&%
Þ!")*
"(
&
Þ"&'$
Þ!$()
")
'
Þ"!%#
Þ!'$"
"*
(
Þ!&*&
Þ!*!"
#!
)
Þ!#*)
Þ""#'
#"
*
Þ!"$#
Þ"#&"
##
"!
Þ!!&$
Þ"#&"
#$
""
Þ!!"*
Þ""$(
#%
"#
Þ!!!'
Þ!*%)
#&
Þ!#"(
Þ!"#)
Þ!!("
Þ!!$(
Þ!!"*
Þ!!!*
Þ!!!%
Þ!!!#
Þ!!!"
Þ!!!!
57. Suppose that \ is exponential with mean -•" Þ Set up the two equations that will
have to be solved simultaneously in order to find UMP size ! test of H! À - Ÿ Þ& or
- 2 vs H" À - - ÐÞ&ß #Ñ.
58. Consider the situation of problem 23. Find the form of UMP tests of H! À ) Ÿ )! vs
H" À ) ž )! Þ Discuss how you would go about choosing an appropriate constant to
produce a test of approximately size !. Discuss how you would go about finding an
optimal (approximately) 90% confidence set for ).
59. (Problem 37, page 122 TSH.) Let \" ß ÞÞÞß \8 and ]" ß ÞÞÞß ]7 be independent samples
from NÐ0ß "Ñ and NÐ(ß "Ñ, and consider the hypotheses H! À ( Ÿ 0 and H" À ( ž 0Þ Show
–
–
that there exists a UMP test of size !, and it rejects H! when ] • \ is too large. (If
0" • (" is a particular alternative, the distribution assigning probability 1 to the point
with ( œ 0 œ Ð70" € 8(" ÑÎÐ7 € 8Ñ is least favorable at size !.)
60. Problem 57 Schervish, page 293.
15
Download