Chapters: 1. Unbiased Estimator 2. Mean Square Error 3. Sufficient Statistics 4. Complete Statistics 5. Exponential family 1. Unbiasedness Definition 1.1: - A statistics π = π(π1 , π2 , … … , ππ ) is called an unbiased estimator of π if πΈ(π) = π. If πΈ(π) ≠ π then T is called a biased estimator of π and bias is defined as, π΅πππ (πΜ) = πΈ(π − π) Definition 1.2: - An estimator π = π(π1 , π2 , … … , ππ ) is called asymptotically unbiased estimator of π if lim πΈ(π) = π or, πβΆ∞ lim π΅πππ (πΜ) = 0 πβΆ∞ Definition 1.3: - The mean square error (MSE) of an estimator πΜ of a parameter 2 π is the function of π defined by πΈ(πΜ − π) , and this is denoted as πππΈπΜ . 2 2 2 πππΈπΜ = πΈ(πΜ − π) = π(πΜ ) + (πΈ(πΜ) − π) = π(πΜ) + (π΅πππ ππ πΜ) 2 2 If an estimator is unbiased then we have, πππΈπΜ = πΈ(πΜ − π) = πΈ (πΜ − πΈ(π)) = π(πΜ) Definition 1.4: - Let T and T’ be two unbiased estimators of the population parameter π, such that πΈ(π) = π and πΈ(π′) = π. Then the estimator T is efficient relative to the estimator T’ if the variance of the finite-sample distribution of T is less than or at most equal to the variance of the finite-sample distribution of T’; That is, if π(π) ≤ π(π′) for all finite n where πΈ(π) = π = πΈ(π′) o Both the estimators T and T’ must be unbiased, since the efficiency property refers only to the variances of unbiased estimators. 1 o An efficiency, πθ of an estimator T* for estimating π is defined as π(π ∗). Example 1.1:- If T is an unbiased estimator of π. Then show that π π is a biased estimator of ππ . Sol: It is given that πΈ(π) = π. We know that, π(π) > 0 βΉ πΈ(π 2 ) − πΈ 2 (π) > 0 βΉ πΈ(π 2 ) − π 2 > 0 [β΅ πΈ(π) = π] βΉ πΈ(π 2 ) > π 2 This proves that π 2 is a biased estimator of π 2 . Example 1.2:- If π π , π π , … . . , π π§ ~πππ«(π©). Then find an unbiased estimator of π©π based on π π , π π , … . . , π π§ . Sol: We know that if π1 , π2 , … . . , ππ ~π΅ππ(π) then πΈ(ππ ) = ππ. Consider a simple statistics ‘T’ based on π1 , π2 , … . . , ππ as π = ∑ππ=1 X i then π~π΅ππ(π, π) and hence πΈ(π) = ππ πππ π(π) = πππ Now, πΈ(π 2 ) − πΈ 2 (π) = πππ βΉ πΈ(π 2 ) − (ππ)2 = πππ = ππ − ππ2 βΉ πΈ(π 2 ) = (ππ)2 + ππ − ππ2 = ππ2 (π − 1) + ππ βΉ πΈ(π 2 ) − ππ = ππ2 (π − 1) βΉ πΈ(π 2 ) − πΈ(π) = ππ2 (π − 1) βΉ πΈ(π 2 − π) = ππ2 (π − 1) π(π−1) βΉ πΈ [π(π−1)] = π2 ∴ π(π−1) π(π−1) is an unbiased estimator of π2 . Example 1.3:- If π π , π π , … . . , π π§ are a random sample from the uniform distribution i.e. π(π, π). Show that [ π§+π π§ π (π§) ] is an unbiased estimator of π. Sol: Given that π1 , π2 , … . . , ππ ~π(0, π). Transform Yi = Xi Therefore, πΈ(π(π) ) = π+1 θ then Y1, Y2, Y3………… Yn ~U(0,1) and Y(j) ~Beta(j, n − j + 1) π π(π) βΉπΈ( π π [β΅ π(π) = ) = π+1 π(π) π π βΉ πΈ(π(π) ) = π+1 π π βΉ πΈ(π(π) ) = π+1 π π+1 βΉπΈ( This proves that π n+1 n π(π) ) = π X (n) is an unbiased estimator of θ. Note: There are many other ways to this. One of them is to first find the pdf of X(n) , where π(π) = πππ₯(π1 , π2 , … . . , ππ ) and compute the expectation. Example 1.4:- If X1 , X2 , … . . , Xn ~Poi(λ). Then find an unbiased estimator of π−λ based on T = ∑ni=1 Xi . Sol: We know that if π1 , π2 , … . . , ππ ~πππ(π), then π = ∑ni=1 Xi ~πππ(ππ). Let β(π) be an unbiased estimator of e−λ Therefore we must have, E(β(π)) = π−λ βΉ −nλ ∑∞ π‘=0 β(π‘)π βΉ π−nλ ∑∞ π‘=0 β(π‘) βΉ ∑∞ π‘=0 β(π‘) By equating we have, βΉ (nλ)t t! (nλ)t t! (nλ)t t! = π−λ = π−λ = πnλ π−λ = π(n−1)λ = ∑∞ π‘=0 β(π‘)nt = (n − 1)t [(n−1)λ]t t! ] βΉ β(π‘) = n−1 t (n−1)t =( nt n ) n−1 T Hence, ( n ) is an unbiased estimator of e−λ . Example 1.5:- Suppose π π , π π , … . . , π π§ are i.i.d. random variables with probability density function 1 f(x|σ) = 2σ exp (− ∑ |π| Show that an estimator π Μ= π§ |x| σ is unbiased. ∞ 1 ) ; −∞ < π₯ < ∞; π > 0 Sol: We have πΈ(|π|) = 2π ∫−∞ |π₯| ππ₯π (− ∞ 2 |π₯| π ) ππ₯ π₯ = 2π ∫0 π₯ ππ₯π (− π) ππ₯ ∞ 1 π₯ = π ∫0 π₯ ππ₯π (− π) ππ₯ [Having an even function] [Mean of exponential dist] =π Hence, πΈ(πΜ) = πΈ ( ∑π π=1 |ππ | π This shows that πΜ = 1 ) = π ∑ππ=1 πΈ(|ππ |) = π ∑π π=1 |ππ | π is an unbiased estimator of π. Example 1.6:- Let π π , π π , … . . , π π§ be a random sample from a Bernoulli distribution with parameter π©; π ≤ π© ≤ π. Find the bias of the estimator estimating π©. for [JAM 2006] Sol: - Since Xi ~ Ber(p); i = 1, 2, … . , n Therefore, ∑ Xi ~Bin(n, p); Now, π§ √π§+π ∑π’=π π π’ π(π§+√π§) n √n+2 ∑i=1 Xi ) 2(n+√n) E( βΉ E(∑ Xi ) = np √n+2np √n) = 2(n+ Hence, Bias = π √π+2 ∑π=1 ππ 2(π+√π) πΈ( √π+2ππ − √π) − π) = 2(π+ √π(1−2π) √π(1+√π) π=2 1 = (1+ √ 1 ( − π) π) 2 Example 1.7:- Let X1 , X2 , … . . , Xn be a random sample from Ber(p); π ≤ π© ≤ π. Show 1 that there doesn’t exists any unbiased estimator of p. Sol: We know that if π1 , π2 , … . . , ππ ~π΅ππ(π) then π = ∑ππ=1 ππ ~π΅ππ(π, π). Let β(π) be 1 a function of T which is an unbiased estimator of . π Therefore we have, 1 πΈ[β(π)] = ∑ππ‘=1 β(π‘)(ππ‘)ππ‘ (1 − π)π−π‘ = π βΉ ∑ππ‘=1 β(π‘)(ππ‘)ππ‘+1 (1 − π)π−π‘ − 1 = 0 In the left hand side, we have a polynomial of degree n+1 in the variable p and it must be zero for all p ∈ (0, 1). That is the polynomial must be identically equal to zero and hence we must have −1 ≡ 0 which is a contradiction. This shows that there doesn’t exists any unbiased estimator of 1 π based on T. ππ Μ ) = , where Example 1.8:- For a random variable X, we have π(π) = π, π > 0 ππ§π π―ππ«(π ππ Μ is an estimate of π and π Μ = π€ π. Also it is given that πππ(π) = π[ππ’ππ¬(π)]π . Find k. π π€+π Sol: π π E(πΜ ) = πΈ (π+1 π) = π+1 π as πΈ(π) = π π −π ∴ π΅πππ = πΈ(πΜ − π) = π+1 π − π = π+1 and πππΈπΜ = π(πΜ) + (π΅πππ ππ πΜ) 2 by definition 2 2 2 −π 2π It is also given that πππΈπΜ = 2[ππππ (πΜ)] = 2 (π+1) = (π+1)2 2 2 Hence π(πΜ) + (π΅πππ ππ πΜ) = πππΈπΜ = 2(π΅πππ ππ πΜ ) 2 βΉ π(πΜ) = (π΅πππ ππ πΜ ) βΉ π2 −π 2 = (π+1) 25 βΉ π=4 Example 1.9:- Let X and Y are random variables with finite means. Find the π function π (π), for which π [(π − π (π)) ] is minimum, where π (π) ranges over all functions. Sol: - WE have, 2 πΈ(π − π(π)) = πΈ(π − πΈ(π|π) + πΈ(π|π) − π(π)) 2 2 2 = πΈ(π − πΈ(π|π)) + πΈ(πΈ(π|π) − π(π)) +2πΈ[(π − πΈ(π|π))(πΈ(π|π) − π(π))] The product term can be shown to zero by iterating the expectation. Thus 2 2 2 πΈ(π − π(π)) = πΈ(π − πΈ(π|π)) + πΈ(πΈ(π|π) − π(π)) 2 ≥ πΈ(π − πΈ(π|π)) , for all π(. ) 2 Therefore, πΈ(π − π(π)) is minimum when π(π) = πΈ(π|π). Example 1.10:- Let π π , π π , … . . , π π§ be a random sample from a π(π, ππ ) distribution, where both µ and ππ are unknown. Find the value of b that minimizes the mean π Μ )π for estimating ππ , where π Μ = squared error of the estimator ππ = π§−π ∑π§π’=π(π π’ − π π π§ ∑π§π’=π π π’ . [JAM 2006] b Sol: Given that, ππ = n−1 ∑ni=1(ππ − πΜ )2 = bπ 2 βΉ πΈ(ππ ) = E(bπ 2 ) = bπ 2 and π(ππ ) = V(bπ 2 ) = Also, πππΈ(ππ ) = π(ππ ) + [π΅πππ (ππ )]2 = 2π 2 π4 n−1 2π 2 π4 n−1 . 2π 2 + (b − 1)2 π 4 = π 4 {π 2 − 2π + 1 + n−1} Therefore, for minimum/maximum we have π ππ πππΈ(ππ ) = 0 4π βΉ π 4 {2π − 2 + n−1} = 0 βΉ 2π(n − 1) − 2(n − 1) + 4b = 0 βΉ 2π(n − 1 + 2) = 2(n − 1) βΉ π = (n+1) (n−1) (n−1) Hence, πππΈ(ππ ) is minimum when π = (n+1). Example 1.11:- Let π π , π π , … . . , π π§ be a random sample from a π(π) distribution. Give two different estimates of π and determine which estimate is better. Sol: Let πΜ and π 2 be the sample mean and sample variance, respectively. 1 Therefore, πΈ(πΜ ) = π and πΈ(π 2 ) = E {n−1 ∑ni=1(ππ − πΜ )2 } = π 2 = λ. 2 4 2 π π 2π 2λ Also, π(πΜ ) = π = π and π(π 2 ) = π−1 = π−1. Hence for some value of π, πΜ is better, π 2 is better. [prove this] 2. Sufficient Statistics A sufficient statistic is one that contains all the necessary information about the unknown parameters in a given population. Definition 2.1: A statistic T(X) is sufficient for a given population parameter π if and only if the distribution of the data π1 , π2 , … . . , ππ given π = π‘ does not depend on the unknown parameter π. Example 2.1:- Let X1 , X2 , … . . , Xn ~Poi(λ). Find sufficient statistics for π. Sol: Let, π = ∑ Xi βΉ π~πππ(ππ) Therefore, P(π1 = π₯1 , π2 = π₯2 , … . , ππ = π₯π |π = π‘) = = = = = P(π1 =π₯1 ,π2 =π₯2 ,….,ππ =π₯π ,π=π‘) P(π=π‘) P(π1 =π₯1 ,π2 =π₯2 ,….,ππ =π₯π ) when ∑ xi = t P(π=π‘) P(π1 =π₯1 ) π(π2 =π₯2 ) …. π(ππ =π₯π ) P(π=π‘) because each Xi′ s are independent e−λ (λ)π₯1 e−λ (λ)π₯2 ,…….e−λ (λ)π₯π π₯1 !π₯2 !……π₯π ! e−nλ (nλ)t e−nλ (λ)∑ xi π₯1 !π₯2 !……π₯π ! e−nλ (nλ)t t! t! =π₯ n−t t! , 1 !π₯2 !……π₯π ! independent of . As, the conditional distribution π1 , π2 , … . . , ππ |π = π‘ is independent of π, hence π = ∑ni=1 Xi is sufficient statistics for π. Example 2.2:- Let X1, X 2 ~Poi(λ). Then show that T = X1, +2X2 is not sufficient statistics for π. Sol: π(π = 2) = π(π1 +2π2 = 2) = π(π1 = 0, π2 = 1) + π(π1 = 2, π2 = 0) = π(π1 = 0)π(π2 = 1) + π(π1 = 2)π(π2 = 0) λ2 = e−λ . λe−λ + e−λ 2! . e−λ λ = λe−2λ (1 + 2) Now, P(π1 = π₯1 , π2 = π₯2 |π = π‘) = = = P(π1 =0,π2 =1,π=2) P(π=2) e−λ .λe−λ λ λe−2λ (1+ ) 2 = P(π1 =π₯1 ,π2 =π₯2 ,π=π‘) P(π1 =0,π2 =1) P(π=π‘) P(π=π‘) = P(π1 =0)π(π2 =1) P(π=π‘) λ not independent of λ = (1 + 2), As, the conditional distribution π1 , π2 |π = π‘ is not independent of π, hence π = π1, +2π2 is not sufficient for λ. Note: To prove not a sufficient statistics, you just need to give only 1 counter example. Example 2.3:- If X1, X2 ~N(µ, 1). Then show that T = X1 +2X2 is not sufficient for µ. Sol: Here, we can’t compute P(T=t) because we are dealing with continuous distribution. However if we can show that the distribution of π1 |π depends on µ, then we can say that T is not sufficient statistics for µ. as π1, π2 ~π(µ, 1) βΉ πΈ(π1 ) = πΈ(π2 ) = µ and π(π1 ) = π(π2 ) = 1 Now, πΈ(π) = πΈ(π1 + 2π2 ) = πΈ(π1 ) + 2πΈ(π2 ) = 3µ π(π) = π(π1 + 2π2 ) = π(π1 ) + 4π(π2 ) + 4πΆππ£(π1 , π2 ) = 5 πΆππ£(π1 , π) = πΆππ£(π1 , π1 +2π2 ) = π(π1 ) + 2πΆππ£(π1 , π2 ) = π(π1 ) = 1 Hence, π= πΆππ£(π,π1 ) √π(π)π(π1 ) = 1 √5 Therefore, (π1 , π)~π΅π (µ, 3µ, 1,5, 1 ) √5 and hence, (π1 |π)~π (µ + 1 1 √5 √5 1 (π − 3µ), (1 − )) 5 That is, (π1 |π)~π ( π+2µ 4 5 , 5) As, the conditional distribution π1 |π = π‘ is not independent of µ, hence π = π1, +2π2 is not sufficient for µ. The method we learn above involves computation of conditional probability and sometimes it may be much difficult to compute the conditional probability. So another method is to compute the sufficient statistics is given by Neyman. The Neyman Factorization Theorem Definition 2.2: A real valued statistic π = π(π1 , π2 , … . . , ππ ) is sufficient for the unknown parameter π if and only if the following factorization holds: πΏ(π) = π(π(π₯1 , π₯2 , … . . , π₯π ); π)β(π₯1 , π₯2 , … . . , π₯π ) for all π₯1 , π₯2 , … . . , π₯π ; where the two functions π(. ; π) and β(. ) are both nonnegative, β(π₯1 , π₯2 , … . . , π₯π ) is independent of π, and π(π(π₯1 , π₯2 , … . . , π₯π ); π) depends on π₯1 , π₯2 , … . . , π₯π only through the observed value π(π₯1 , π₯2 , … . . , π₯π ) of the statistic T with π. Note: o In the above statement it is not mandatory that π(π(π₯1 , π₯2 , … . . , π₯π ); π) must be the pmf or the pdf of π(π₯1 , π₯2 , … . . , π₯π ). However, it is essential that the function β(π₯1 , π₯2 , … . . , π₯π ) must be entirely free from π. o It should be noted that the splitting of likelihood function πΏ(π) may not be unique, there may be more than one way to determine the function β(. ) so that the above relation holds. That’s why there can be different versions of the sufficient statistics. Example 2.4: If X1 , X2 , … . . , Xn ~Bin(p). Then find sufficient statistics for p. Sol: Here, πΏ(π) = ∏ππ=1 π π₯π (1 − π)1−π₯π = π∑ π₯π (1 − π)π−∑ π₯π , Now consider π(∑ππ=1 π₯π ; π) = π∑ π₯π (1 − π)π−∑ π₯π πππ β(π₯1 , π₯2 , … , π₯π ) = 1 Therefore, πΏ(π) = π(∑ππ=1 π₯π for all π₯1 , π₯2 , … , π₯π ∈ {0,1} ; π)β(π₯1 , π₯2 , … , π₯π ) Hence π = π(π1 , π2 , … . . , ππ ) = ∑ππ=1 ππ is sufficient for p. Example 2.5:- If π π , π π , … . . , π π§ ~π(µ, ππ ) where µ and ππ are both unknown. Find the joint sufficient statistics for π = (µ, π). Sol: Here, πΏ(π) = {π√2π} −π 1 ππ₯π {− 2π2 (∑ππ=1 π₯π2 − 2µ ∑ππ=1 π₯π + πµ2 )}, 1 Now consider, π(∑ππ=1 π₯π , ∑ππ=1 π₯π2 ; π) = π −π ππ₯π {− 2π2 (∑ππ=1 π₯π2 − 2µ ∑ππ=1 π₯π + πµ2 )}, and β(π₯1, π₯2, … . . , π₯π ) = {√2π} −π for all π₯1, π₯2, … . . , π₯π ∈ π π Therefore, πΏ(π) = π(∑ππ=1 π₯π , ∑ππ=1 π₯π2 ; π)β(π₯1, π₯2, … . . , π₯π ) Hence π = π(π1 , π2 , … . . , ππ ) = (∑ππ=1 ππ , ∑ππ=1 ππ2 ) is jointly sufficient for (µ, π). Note: o The sample itself π = (π1 , … … . . , ππ ) is always sufficient for π, but this provides no data reduction. o The order statistics π(1) , … … . . , π(π) are always sufficient for π. o If T is a sufficient statistic for π, then π(π), which is a one-to-one function of T, is also sufficient for π. Μ = 1 ∑ππ=1 ππ , π 2 = 1 {∑ππ=1 ππ2 − o So in above example if we consider,π 1 (∑ππ=1 ππ )2 }, π (πΜ 2 ) π then it is one-to-one function of π = π−1 π (∑π=1 ππ , ∑ππ=1 ππ2 ). Hence π = , π is also jointly sufficient for(µ, π 2 ). o If π = (π1 , π2 ) is jointly sufficient for π = (π1 , π2 ). Then it doesn’t imply that π1 is sufficient for π1 or π2 is sufficient for π2 . So in above example we can’t say that alone πΜ is sufficient for µ and π 2 is sufficient for π 2 because πΜ has some information about both µ and π 2 , whereas π 2 has information about π 2 alone. o If T be a sufficient statistic for π then any arbitrary function T’ of T is not necessarily sufficient for π. Example 2.6:- Suppose the statistic T is sufficient for π. Then any monotonic function π(π) will also be sufficient for π. Sol: Let π = π(π). Since π(π) is monotonic therefore π = π −1 (π) exists. Now, for all π₯1 , π₯2 , … . . , π₯π πΏ(π) = π(π; π)β(π₯1 , π₯2 , … . . , π₯π ) πΏ(π) = π(π −1 (π); π)β(π₯1 , π₯2 , … . . , π₯π ) πΏ(π) = π(π; π)β(π₯1 , π₯2 , … . . , π₯π ) Hence, Y is sufficient for π. Example 2.7:- Suppose that X is distributed as π(π, π) where −∞ < π < ∞ is the unknown parameter. Obviously, T = X is sufficient for π. One should check that the statistic T’ = |X|, a function of T, is not sufficient for π. Sol: Let πΊ(. ) be the distribution function of T=|X|. Therefore, πΊ(π‘) = π(|π| ≤ π‘) = π(−π‘ ≤ π ≤ π‘) = π·(π‘ − π) − π·(−π‘ − π) πΉ(π₯, π‘) = π(π ≤ π₯, π ≤ π‘) = π(π ≤ π₯, |π| ≤ π‘) = π(∞ < π ≤ π₯, −π‘ ≤ π ≤ π‘) = π{πππ₯(∞, −π‘) < π < πππ(π₯, π‘)} = π{−π‘ < π < πππ(π₯, π‘)} = π·{πππ(π₯, π‘) − π} − π·(−π‘ − π). π(π₯, π‘) = Example 2.8:- Suppose that π π , π π , … . , π π§ be a random sample from π(π, π) distributed, where π(> 0) is unknown. Find sufficient statistics for π. Sol: Here, 1 π(π₯π ) = π πΌ(0 < π₯π < π), where I is an indicator function defines as πΌ = { 1, π₯ < π 0, ππ‘βπππ€ππ π 1 1 πΏ(π) = ∏ππ=1 π πΌ(0 < π₯π < π) = ππ πΌ(0 < π₯(π) < π) 1 Now consider, π(π₯(π) ; π) = ππ πΌ(0 < π₯(π) < π) πππ β(π₯1 , π₯2 , … , π₯π ) = 1 for all π₯1 , π₯2 , … , π₯π ∈ (0, π) Therefore, πΏ(π) = π(π₯(π) ; π)β(π₯1 , π₯2 , … , π₯π ) Hence π = π(π1 , π2 , … . , ππ ) = π(π) is sufficient for π. Example 2.9:- Let π π , π π , π π be independent random variables with π π€ (k = 1,2,3) having the probability density function −kθx , fk (x) = {kθe 0, 0<π₯<∞ ππ‘βπππ€ππ π where π > π. Then find a sufficient statistics for π. [JAM 2008] Sol: The likelihood function is given by πΏ(π) = π1 (x1 )π2 (x2 )π3 (x3 ) = πe−θx1 . 2πe−2θx2 . 3πe−3θx3 , 0< π₯1, x2 , x3 < ∞ = 6π 3 e−θ(x1+2x2+3x3) This shows that x1 + 2x2 + 3x3 is sufficient for θ. Note: o The above example illustrate that we can also used Neyman Factorization Theorem to prove our sufficient statistics for non-iid case. Example 2.10:- The distribution of π = (X1 , X 2 , X3 , X4 ) is given by 1 θ P(X1 = x1 ) = 2 + 4 , P(X2 = x2 ) = P(X3 = x3 ) = Show that (X1 , X2 + X3 , X4 ) is sufficient statistics for π. Sol: The joint distribution of (X1 , X2 , X3 , X4 ) is 1−θ 4 θ , P(X4 = x4 ) = 4 P(X = x) = x =x n! 1 1 !x2 !x3 !x4 n! 1 !x2 !x3 !x4 θ x1 ( + 4) ( ! 2 1 θ x1 ( + 4) ( ! 2 1−θ x2 1−θ x3 ) ( 4 1−θ x2 +x3 4 ) 4 θ x4 ) (4 ) θ x4 (4 ) = π(x1 , x2 + x3 , x4 , θ)β(π₯1 , π₯2 , π₯3 , π₯π ) Therefore, π = (X1 , X2 + X3 , X4 ) is sufficient statistics for π. Example 2.11:- Let the probability mass function of a random variable X be P(X = 0) = θ, P(X = 1) = 2θ, P(X = 2) = 1 − 3θ; 1 0 < π < 3. Let (ππ , ππ , ππ ) be the vector of observed frequency of 0,1,2 in a sample of size n. Show that (ππ , ππ , ππ ) is a sufficient statistics. Further assume that ππ = ππ + ππ . Then using the result that T4 ~Bin(n, 3θ), show that ππ is also sufficient for π. Sol: The joint distribution of (X1 , X2 , … … , Xn ) is P(X = x) = (θ)t1 (2θ)t2 (1 − 3θ)t3 [where t1 + t 2 + t 3 = n] Now consider, π(t1 , t 2 , t 3 , θ) = (θ)t1 (2θ)t2 (1 − 3θ)t3 πππ β(π₯1 , π₯2 , … , π₯π ) = 1 for all π₯1 , π₯2 , … , π₯π Therefore, P(X = x) = π(t1 , t 2 , t 3 , θ)β(π₯1 , π₯2 , … . , π₯π ) Hence, π = (T1 , T2 , T3 ) is sufficient statistics for π. Also, P(X = x|T4 ) = π(π1 = π₯1 , … . , ππ = π₯π |T4 = t) = = π(π1 =π₯1 ,….,ππ =π₯π ) π(T4 =t) (θ)t1 (2θ)t2 (1−3θ)t3 (nt)(3θ)t (1−3θ)n−t 2t2 (θ)t1 +t2 (1−3θ)n−t1 +t2 = (n)(3θ)t1+t2 (1−3θ)n−t1+t2 t 2t2 = (n)3t t [as π‘ = t1 + t 2 and t1 + t 2 + t 3 = n] As the conditional distribution of π|T4 is independent of π, hence T4 is also sufficient for π. Example 2.12:- Let X1, X2, … … , Xn be a random sample from a distribution with density π1 π(π₯|π1 π2 ) = { π −π₯/π2 , π₯ > 0 π2 π1 (1 − π ) π π₯/π2 , π₯ < 0 , 2 where 0 < π1 < 1 and π2 > 0. Show that (πΊ, π²) is sufficient for (π½π , π½π ), where K is the number of positive πΏπ ′π, π² = ∑ππ=π π°(πΏπ > π), and πΊ = ∑ππ=π |πΏπ |. Sol: Let k πΏπ ′π, out of n, are positive and n-k πΏπ ′π are negative. The likelihood function is given by π π π π(π₯|π1 π2 ) = [π1 π −π₯/π2 ] [(1 − π1 ) π π₯/π2 ] 2 π π−π 2 π π π(π₯|π1 π2 ) = (π1 ) (1 − π1 ) 2 2 = π1 π (π2 −π1 )π−π π2 π π−π π − π − ∑1 π₯/π2 π ∑2 π₯/π2 1 (∑ π₯−∑2 π₯) π2 1 , where ∑1 π₯ is the sum of all ππ ′π which are positive and ∑2 π₯ is the sum of all ππ ′π which are negative. = π1 π (π2 −π1 )π−π π2 π π − 1 (∑π |πΏ |) π2 π=π π , [ We note that ∑1 π₯ − ∑2 π₯ = ∑ππ=π |πΏπ |; Suppose X=-10, -5, 2, 3, 5, then ∑1 π₯ = 2 + 3 + 5 = 10 and ∑2 π₯ = −10 − 5 = −15 and ∑ππ=π |πΏπ | = ππ + π + π + π + π = ππ. That is ∑1 π₯ − ∑2 π₯ = 10 − (−15) = 25 = ∑ππ=π |πΏπ | ] Now consider, π(π, πΎ; π1 , π2 ) = π1 π (π2 −π1 )π−π π2 π π − 1 (∑π |πΏ |) π2 π=π π πππ β(π₯1 , π₯2 , … , π₯π ) = 1 for all π₯1 , π₯2 , … , π₯π ∈ (0, π) Therefore, πΏ(π) = π(π, πΎ; π1 , π2 )β(π₯1 , π₯2 , … , π₯π ) Hence π = π(π1 , π2 , … . , ππ ) = (πΊ, π²) is sufficient for (π½π , π½π ). 3. Ancillary Statistics We have learned that a sufficient statistic T preserves all the information about π contained in the data X. However, an ancillary statistic T by itself provides no information about the unknown parameter π. But still we learn them because they can play an important role in statistical methodologies. Definition 3.1: A statistic T is called ancillary for π provided that the pmf or pdf of T, denoted by f(t) for t ∈ T, does not involve the unknown parameter π ∈ Θ. Example 3.1: Suppose that πΏπ, πΏπ, πΏπ………… πΏπ are iid π΅(π½, π) where π ∈ πΉπ is unknown. Find few ancillary statistics. Sol: (1) A statistic π1 = π1 −π2 is distributed as π(0,2) whatever be the value of the unknown parameter π. Hence π1 is ancillary for π. (2) Another statistic π2 = π1 +π2 + β― +ππ−1 − (π − 1)ππ is distributed as π(0, π(π − 1)) whatever be the value of the unknown parameter π. Hence π2 is also ancillary for π. 1 2 (3) The sample variance π 2 is distributed as π−1 ππ−1 whatever be the value of the unknown parameter π and hence π 2 is ancillary too for π. Example 3.2: Suppose that πΏπ, πΏπ, πΏπ………… πΏπ are iid πππ(π½), where π (>0) is the unknown parameter with n = 2. Find ancillary statistics. Sol: Define ππ = πππ for π = 1, … . π, then the joint distribution of π = π1, π2………… ππ does not involve the unknown parameter π. π π Define π1 = π1 = π1 . We can observe that the distribution of π1 is independent of π π π and so π1 is ancillary. π2 Also, define the statistic π2 = ∑1 π=1 ππ π2 = ∑1 π=1 ππ and we can observe that the distribution of π2 is independent of π and so π2 is ancillary. Note: o Defining Ζ1 , Ζ2 πππ Ζ3 be the location, scale and location-scale family defines as Ζ1 = {π(π₯; π) = π(π₯ − π): π ∈ R, π₯ ∈ R }; π₯ Ζ2 = {π(π₯; πΏ) = πΏ −1 π (πΏ) : πΏ ∈ R+ , π₯ ∈ R } ; Ζ3 = {π(π₯; π, πΏ) = πΏ −1 π ( π₯−π πΏ ) : π ∈ R, πΏ ∈ R+ , π₯ ∈ R } ; where π, πΏ may belong to some appropriate subsets of R and R+ o Let the common pdf of π1, π2, π3………… ππ belongs to the location family Ζ1 . Then the statistics U = (π1 − ππ , π2 − ππ , … … , ππ−1 − ππ ) is ancillary. o Let the common pdf of π1, π2, π3………… ππ belongs to the scale family Ζ2 . Then π π the statistics V = ( π1 , π2 , … … , π π ππ−1 ππ ) is ancillary. o Let the common pdf of π1, π2, π3………… ππ belongs to the location-scale family π1 −ππ π2 −ππ Ζ3 . Then the statistics V = ( is the sample variance. π , π ,……, ππ−1 −ππ π ) is ancillary, where S2 Complete Statistics 4. Definition 4.1: - A statistic T, is called complete if and only if for any real valued function π(π‘) defined for t ∈ T, having finite expectation, such that πΈ[π(π‘)] = 0 for all π ∈ Θ implies π[π(π‘) = 0] = π. Note: o o o o If If If If a non-constant function of T is ancillary, then T(X) is not complete. E(T) does not depend on π, then T(X) is not complete. T is complete then (T), is also complete. T is complete, then only one unbiased estimator based on T is possible. Example 4.1: Suppose that a statistic T is distributed as π΅ππ(π), 0 < π < 1. Let us examine if T is a complete statistic. Sol: The pmf of T is given by π(π‘; π) = ππ‘ (1 − π)1−π‘ , π‘ = 0, 1 πππ 0 < π < 1. Consider any real valued function β(π‘) such that πΈ[β(π‘)] = 0 for all 0 < π < 1. πΈ[β(π‘)] = ∑1π‘=0 β(π‘) π(π‘; π) = β(0)(1 − π) + β(1)π = 0 βΉ β(0) + [β(0) − β(1)]π = 0 As above equation is linear in p and so it may be zero for exactly one value of p between 0 and 1. But we need that β(0) + [β(0) − β(1)]π must be zero for infinitely many values of p in (0, 1). Hence, this expression must be identically equal to zero which means that β(0) = 0 and β(0) − β(1) = 0. βΉ β(0) = 0 πππ β(1) = 0 In other words, we have β(π‘) = 0 for π‘ = 0, 1. Thus,T is a complete statistic. Definition 4.2: A statistic T is called complete sufficient for an unknown parameter π if and only if (i) T is sufficient for π and (ii) T is complete. Problem 4.2: Suppose that π π, π π, π π………… π π§ be iid Poi(λ), λ > 0. Show that π = ∑ni=1 Xi is a complete statistic. Sol: We already have shown that T = ∑ni=1 Xi is a sufficient statistics for π. So need to prove that T is also complete. We also know that T = ∑ni=1 X i ~πππ(ππ). Consider any real valued function β(π‘) such that πΈ[β(π‘)] = 0 for all π > 0. ∞ −ππ πΈ[β(π‘)] = ∑∞ π‘=0 β(π‘) π(π‘; π) = ∑π‘=0 β(π‘) π (ππ)π‘ π‘! ∀ π‘ ∈ π = {0,1, … } = 0; ππ‘ π‘ βΉ ∑∞ π‘=0 π(π‘) π = 0, where π(π‘) = β(π‘) π‘! We can see that the above expression can be expressed in an infinite power of π (> 0) and if the infinite power series is equal to zero then each term will be equal to zero. Hence, π(π‘) = 0 ∀ π‘ = 0,1, … βΉ β(π‘) ≡ 0 ∀ π‘ = 0,1, … Therefore, T is complete statistics. Example 4.3: Suppose that X1, X2, X3………… Xn be iid N(θ, θ2 ), θ > 0. Show that π = Μ , π π ), is jointly sufficient for π but not complete. (π Sol: Here, πΏ(π) = {√2πθ2 } −π 1 ππ₯π {− 2θ2 (∑ππ=1 π₯π2 − 2π ∑ππ=1 π₯π + ππ 2 )}, 1 Now consider, π(∑ππ=1 π₯π , ∑ππ=1 π₯π2 ; π) = π −π/2 ππ₯π {− 2θ2 (∑ππ=1 π₯π2 − 2π ∑ππ=1 π₯π + ππ 2 )}, and β(π₯1, π₯2, … . . , π₯π ) = {√2π} −π for all π₯1, π₯2, … . . , π₯π ∈ π π Therefore, πΏ(π) = π(∑ππ=1 π₯π , ∑ππ=1 π₯π2 ; π)β(π₯1, π₯2, … . . , π₯π ) Hence π′ = π′(π1 , π2 , … . . , ππ ) = (∑ππ=1 ππ , ∑ππ=1 ππ2 ) is jointly sufficient for π. Also π = (πΜ , π 2 ) is jointly sufficient for π because it is one-to-one function of π ′ = (∑ππ=1 ππ , ∑ππ=1 ππ2 ). Now aim is to show that ‘T’ is not complete. We know that, πΈ(πΜ ) = π and also πΈ(π 2 ) = θ2 for all π>0. Thus we have, πΈ(πΜ 2 − π 2 ) = 0. Consider β(π) = πΜ 2 − π 2 and then we have πΈ[β(π)] ≡ 0 for all π>0 but β(π‘) = π₯Μ 2 − π 2 is not identically zero ∀ π‘ ∈ π × π + . Therefore, T is jointly sufficient for π but not complete. Example 4.4: Suppose that X1, X2, X3………… Xn be iid with the common pdf given by 1 f(x) = θ exp {− (x−θ) θ }, where x > π Check whether the sufficient statistics is complete or not. Sol: 1 Here, π(π₯) = π ππ₯π {− (π₯−π) π } πΌ(π < π₯ < ∞), where I is an indicator function defined as πΌ = { 1 πΏ(π) = ∏ππ=1 π ππ₯π {− (π₯−π) π 1 } πΌ(π < π₯π < ∞) = ππ ππ₯π {− 1 Now consider π(∑ππ=1 π₯π , π₯(1) ; π) = ππ ππ₯π {− Therefore, πΏ(π) = 1, π₯ > π 0, ππ‘βπππ€ππ π ∑(π₯−π) π } πΌ(π₯(1) > π) ∑(π₯−π) π } πΌ(π₯(1) > π) πππ β(π₯1, π₯2, … . . , π₯π ) = 1 π(∑ππ=1 π₯π , π₯(1) ; π)β(π₯1, π₯2, … . . , π₯π ) for all π₯1, π₯2, … . . , π₯π ∈ (π, ∞) Hence π′ = π′(π1, π2, π3………… ππ ) = (∑ππ=1 ππ , π(1) ) is jointly sufficient for π. Also π = (∑ππ=1(ππ − π(1) ), π(1) ) is one to one function of T’ and hence T is also jointly sufficient for π. Now let π§ = π₯−π π βΉ πππ§ = ππ₯. Also ππ(1) = ππππ(π1 , … ππ ) = πππ(π1 , … ππ ) − π = π(1) − π Therefore, π(π§) = π −π§ ; π§>0 βΉ π~ππ₯π(1) and hence, π(1) −π βΉ πΈ(π(1) ) = πΈ ( π π βΉ πΈ(π(1) ) = π + π = π(1) = πππ(π1 , … ππ )~ππ₯π(π) 1 )=π π(1+π) . π We also have πΈ(∑ππ=1 ππ ) = ∑ππ=1 πΈ(ππ ) = ∑ππ=1 πΈ(π + πππ ) = 2ππ Therefore, πΈ(∑ππ=1(ππ − π(1) )) = 2ππ − π(1 + π) = (π − 1)π βΉ 1 πΈ [(π−1) ∑ππ=1(ππ − π(1) )] = π Let us define, 1 −1 β(π‘) = (1 + ) π βΉ π₯(1) − (π − 1)−1 ∑ππ=1(π₯π − π₯(1) ) 1 −1 πΈ[β(π)] = πΈ [(1 + π) 1 −1 βΉ = (1 + π) βΉ = (1 + π) βΉ =π−π =0 1 −1 ∀ π‘ ∈ π = {π, ∞} × π + π(1) − (π − 1)−1 ∑ππ=1(ππ − π(1) )] 1 πΈ(π(1) ) − π−1 πΈ[∑ππ=1(ππ − π(1) )] 1 πΈ(π(1) ) − π−1 πΈ[∑ππ=1(ππ − π(1) )] 1 −1 Therefore, πΈ[β(π)] ≡ 0 for all π>0 but β(π‘) = (1 + π) π₯(1) − (π − 1)−1 ∑ππ=1(π₯π − π₯(1) ) is not identically zero ∀ π‘ ∈ π = {π, ∞} × π + . Therefore, π = (∑ππ=1(ππ − π(1) ), π(1) ) is sufficient for π but not complete. Example 4.5:- Suppose that π π, π π, π π………… π π§ ~π(π, π) with π > 0 being the unknown parameter. Show that π = π (π§) is complete sufficient statistics for π. Sol: We already have shown that π = π(π) = πππ₯(πΏπ, πΏπ, … … . , πΏπ ) is a sufficient statistic for π. In this case and the pdf of T is given by π(π‘) = π π‘ π−1 ππ ; 0<π‘<π Let β(π‘) be any arbitrary real valued function such that πΈ[β(π‘)] = 0 for all π> 0. Therefore we have, θ E[h(t)] = ∫0 h(t) n tn−1 θn dt = 0 θ βΉ ∫0 h(t) t n−1 dt = 0 βΉ h(θ)θn−1 = 0 [using derivatives] for all θ > 0 βΉ h(θ) ≡ 0, This shows that T = X (n) is complete. Hence π = π(π) is complete sufficient statistics for π. Example 4.6:- Suppose that π π , π π , … … , π π§ are iid with the density function π(x; θ) = e−(x−θ) ; π₯ > θ, Show that π» = πΏ(π) is complete statistic for π. Sol: Let πΉπ (. ) be the distribution function of X, πΉπ (π‘) = 1 − e−(t−θ) Now, ππ (π‘) = πππ (π‘)(1 − πΉπ (π‘)) βΉ 1 − πΉπ (π‘) = e−(t−θ) π−1 = ne−(t−θ) [e−(t−θ) ] π(π‘) = ne−n(t−θ) ; n−1 = ne−n(t−θ) π‘>π Let β(π‘) be any arbitrary real valued function such that πΈ[β(π‘)] = 0 for all π> 0. Therefore we have, ∞ E[h(t)] = ∫θ h(t) ne−n(t−θ) dt = 0 ∞ βΉ ∫θ h(t) e−nt dt = 0 [β΅ ne−nθ > 0] βΉ −h(θ)enθ = 0 This shows that T = X (1) [using derivatives] for all θ > 0 βΉ h(θ) ≡ 0, is complete. Example 4.7:- If there exists an unbiased estimator of π which is complete sufficient statistics, then it is unique. Sol: Suppose T is complete sufficient statistics. Further assume that π1 (T) and π2 (T) be two unbiased estimator of π. Therefore, πΈ[π1 (T)] = π = πΈ[π2 (T)] βΉ πΈ[π1 (T) − π2 (T)] = 0 Hence, π1 (T) ≡ π2 (T) with probability 1. Example 4.8:- Consider the probability mass function of X is given by π(X = x) = (1 − θ)2 θx ; π₯ = 0,1,2,3 …. = θ; π₯ = −1 = 0; otherwise Show that the family is not complete. Sol: Let π(X) be any real valued function such that πΈ[π(X)] = 0. βΉ ∑∞ π₯=−1 π(X)π(X = x) = 0 βΉ π(−1)π(X = −1) + ∑∞ π₯=0 π(X)π(X = x) = 0 βΉ 2 x π(−1)θ + ∑∞ π₯=0 π(X)(1 − θ) θ = 0 βΉ x −2 ∑∞ π₯=0 π(X)θ = − (1−θ)2 π(−1) = −π(−1)θ(1 − θ) βΉ ∞ x 2 x ∑∞ π₯=0 π(X)θ = −π(−1)θ[1 + 2θ + 3θ + β― . . ] = −π(−1) ∑π₯=1 xθ βΉ ∞ ∞ x x x ∑∞ π₯=0 π(X)θ = − ∑π₯=1 π(−1)xθ = ∑π₯=0 −π(−1)xθ θ By equating we have, π(x) = −π₯π(−1), This shows that the family is not complete. Exponential Parameter Family: (1) One parameter form: Suppose that π π, π π, … … , π π§ are iid with the common pmf/pdf belonging to the 1-parameter exponential family if the likelihood is defined by π(π₯; π) = π(π₯)π(π)ππ₯π{π(π)π(π₯)}. Then π(π₯) is complete sufficient statistics for π. (2) Two parameter form: Suppose that X1, X2, … … , Xn are iid with the common pmf/pdf belonging to the 2-parameter exponential family if the likelihood is defined by π(π₯; π) = π(π₯)π(π)ππ₯π{π1 (π1 )π1 (π₯) + π2 (π2 )π2 (π₯)}. Then π = (π1 , π2 ) is complete sufficient statistics for π. Bernoulli, Binomial, Poisson, Normal, Exponential, Gamma, Beta, Geometric etc. are the example of exponential family. o For Bernoulli: π(π₯; π) = π ∑π π=1 π₯π (1 − π) π−∑π π=1 π₯π π π ∑π π=1 π₯π = (1 − π) (1−π) π = (1 − π)π ππ₯π [∑ππ=1 π₯π ππ (1−π)] Therefore, π = ∑ππ=1 ππ is complete sufficient for π. π o For Poisson: π(π₯; π) = π −ππ π₯ π∑π=1 π₯π 1 !π₯2 !……π₯π = π −ππ ! ππ₯π(∑π π=1 π₯π πππ) π₯1 !π₯2 !……π₯π ! Therefore, π = ∑ππ=1 ππ is complete sufficient for π. o For Normal: π(π₯; π) = π₯π −µ 2 1 − (∑π 1 π=1( π ) ) 2 π (2ππ)π/2 = π 2 1 ∑π=1 π₯π π2 − ( 1 2 π (2ππ)π/2 − 2µ ∑π π=1 π₯π π2 Therefore, π = (∑ππ=1 ππ , ∑ππ=1 ππ2 ) is complete sufficient for θ= (µ, π 2 ). π o For Exponential: π(π₯; π) = ππ π −π ∑π=1 π₯π = ππ ππ₯π(π ∑ππ=1 π₯π ) Therefore, π = ∑ππ=1 ππ is complete sufficient for π. πΌπ½ π π½−1 o For Gamma: π(π₯; πΌ, π½) = π€π½ π −πΌ ∑π=1 π₯π ∏ππ=1 π₯π πΌπ½ = π€π½ ππ₯π(−πΌ ∑ππ=1 π₯π + (π½ − 1) ∑ππ=1 πππ₯π ) + πµ2 ) π2 Therefore, π = (∑ππ=1 ππ , ∑ππ=1 ππππ ) is complete sufficient for (πΌ, π½). π o For Geometric: π(π₯; π) = π(1 − π)∑π=1 π₯π = π ππ₯π(∑ππ=1 π₯π ππ(1 − π)) Therefore, π = ∑ππ=1 ππ is complete sufficient for p. Basu’s Theorem: Suppose that we have two statistics, T = T(X) which is complete sufficient for π and W = W(X) which is ancillary for π. Then, T and W are independently distributed. Proof: To prove this theorem, we need to show P(W ≤ w|T = t) = P(W ≤ w). Then we can say that W and T are independently distributed. Since, T is complete sufficient statistics, therefore π(π‘) = P(W ≤ w|T = t) is independent of π and as W is ancillary statistics, P(W ≤ w) is independent of π. Now, let us define 1, W≤w Iw = { 0, π>π€ E[E(Iw |T)] = E(Iw ) = P(W ≤ w) βΉ E[P(W ≤ w|T)] = P(W ≤ w) βΉ E[g(T) − P(W ≤ w)] = 0 As T is complete statistics, so we have if E[Ψ(T)] = 0 then Ψ(T) = 0. Hence g(T) = P(W ≤ w) or P(W ≤ w|T) = P(W ≤ w). Example 4.9:- Let π π, π π, … … , π π§ ~π(µ, ππ ) with n ≥ 2, (µ, σ) ∈ ℜ × ℜ+, where µ is Μ and π π are independently distributed. unknown, but σ is known. Show that, π Sol: We know that π = Μ X is complete sufficient statistic for µ. Also it can be shown that W = S 2 is an ancillary statistic for µ because the distribution of W doesn’t involve µ. Μ and π π are then independently Now, by Basu’s Theorem, the two statistics π distributed. o o o o Μ X and (Xi − Μ X) are also independently distributed. Μ and (Xi − M) are also independently distributed (M is the median). X Μ and (X(n) − X Μ ) are also independently distributed. X (X1 − X2 ) and (X1 + X2 ) are independently distributed because (X1 + X2 ) is sufficient statistics and(X1 − X2 ) is ancillary, when X1 , X2 ~N(µ, σ2 ). 4.10:- Suppose that π π, π π, … … , π π§ ~π(π, π) with n ≥ 2, π (< 0) being the unknown parameter. Show that π (π§) and π (π) π (π§) are independently distributed. Sol: We know that U = X(n) is a complete sufficient statistic for π. It can be X(1) shown that π = X X(1) X(n) (n) which is ancillary for π. Hence, by Basu.s Theorem, X(n) and are independently distributed. X(1) o E(X(1) ) = E (X (n) X(1) X(1) E(X(1) ) (n) (n) (n) ) . X(n) ) = E (X ) E(X(n) ) βΉ E (X ) = E(X Μ X o X(n) and (S ) are also independently distributed. 4.11:- Suppose that π π, π π, … … , π π§ ~π(π, π). Find the covariance between sample mean and sample median. Μ and M be the sample mean and sample median. Then it is seen that X Μ Sol: Let X Μ be complete sufficient statistics for π and (X − M) is ancillary. Then Basu says Μ − M) are independently distributed. that Μ X and (X 1 Μ , X Μ − M) = π(X Μ ) − πππ£(X Μ , M) = − πππ£(X Μ , M) Therefore, 0 = πππ£(X n 1 Μ , M) = βΉ πππ£(X n 4.12:- Let X1, X2, … … , X n ~exp(θ). Define π(X) = π ππ π +π π +β―+π π§ . Find π¬[π(X)]. Sol: Let ππ = πππ , then ππ′ π are exp(1) and hence independent of π. Then we have Z1 π(X) = Z +Z +β―+Z = π(Z) (say). Now π(Z) ~π(X) and hence π(X) is ancillary 1 statistics. 2 n Therefore π = πΈ(X1 ) = E [(X X1 1 +X2 +β―+Xn ) (X1 + X 2 + β― + Xn )] = E[π(X)T(X)] = E[π(X)]E[T(X)] = E[π(X)]nθ Hence, 1 E[π(X)] = π [As T(X) is complete sufficient statistics] [by Basu’s Theorem] [As E[T(X)] = ππ] 5. Consistency An estimator π(= πΜπ ) is said to be a consistent estimator of π if, for all π > 0, lim π(|πΜπ − π| > π) = 0; π→∞ that is, the sequence of real numbers π(|πΜπ − π| > π) → 0 as π → ∞. Consistency is a desirable large-sample property. If πΜπ is consistent, then the probability that the estimator πΜπ differs from the true π becomes small as the sample size n increases. On the other hand, if you have an estimator that is not consistent, then no matter how many data you collect, the estimator πΜπ may never “converge” to π. Definition 5.1: If an estimator πΜπ is consistent, we say that πΜπ converges in π probability to π and write πΜπ βΆ π. Suppose that πΜπ is an estimator of π. If both πΈ(πΜπ ) βΆ θ and π(πΜπ ) βΆ 0, as π βΆ ∞, then πΜπ is a consistent estimator for π. In many problems, it will be much easier to show that πΈ(πΜπ ) βΆ θ and π(πΜπ ) βΆ 0, as π βΆ ∞, rather than showing π(|πΜπ − π| > π) βΆ 0 i.e., appealing directly to the definition of consistency. Example 5.1: Show that the estimator T is consistent for π if πΈ(π) βΆ θ and π(π) βΆ 0, as π βΆ ∞. Sol: For consistency we need to prove π(|π − π| > π) βΆ 0 as π βΆ ∞. We know that, π(|π − π| > π) = π((π − π)2 > π 2 ) ≤ = = πΈ(π−πΈ(π)+πΈ(π)−π)2 π2 π(π)+[πΈ(π)−π]2 π2 πΈ(π−π)2 π2 2 = πΈ(π−πΈ(π)) +[πΈ(π)−π]2 π2 βΆ 0 as π βΆ ∞ because πΈ(π) βΆ θ ⇔ πΈ(π) − π βΆ 0 and also π(π) βΆ 0, as π βΆ ∞. Hence, T is consistent for π. Example 5.2: If T is consistent for π and g is a function continuous at π, then π(π) is consistent for π(π). Sol: For any continuous function g, we have {|π(π(π) ) − π(π)| ≥ π} βΉ {|π(π) − π| ≥ πΏ} Hence, π{|π(π(π) ) − π(π)| ≥ π} ≤ π{|π(π) − π| ≥ πΏ} βΆ 0, as π βΆ ∞ Hence π(π) is consistent for π(π). Example 5.1: Suppose that X1 , … . , Xn is an iid sample from U(0,π), π>0. Show that πΏ(π) is consistent for π. Sol: We have, π(|π(π) − π| > π) = π(π(π) − π > π) + π(π(π) − π < −π) = π(π(π) > π + π) + π(π(π) < π − π) = π(π(π) < π − π) [since π(π(1) > π + π) = 0] = [π(π < π − π)]π [when π > π] π π = (1 − π) βΆ 0, as π βΆ ∞ π [since 0 < (1 − π) < 1] Hence, πΜπ = π(π) is a consistent estimator of π. Example 5.1: Suppose that X1 , … . , Xn is an iid sample from a shifted-exponential distribution −(π₯−π) , π₯>π π(π₯; π) = {π 0, ππ‘βπππ€ππ π. Show that the first order statistic πΜπ = π(1) is a consistent estimator of π. Sol: We have, π(|π(1) − π| > π) = π(π(1) − π > π) + π(π(1) − π < −π) = π(π(1) > π + π) + π(π(1) < π − π) = π(π(1) > π + π) = [π(π > π + π)]π = π −ππ βΆ 0, as π βΆ ∞ [since π(π(1) < π − π) = 0] Hence, πΜπ = π(1) is a consistent estimator of π. Example 5.2:- Let π π , π π , … . . , π π§ be a random sample from a Bernoulli distribution with parameter π©; π ≤ π© ≤ π. Check whether the estimator p is consistent or not. for estimating [JAM 2011] Sol: - Since Xi ~ Ber(p); i = 1, 2, … . , n Therefore, ∑ Xi ~Bin(n, p); n Now, Xi √n+2 ∑ E ( 2(n+ i=1 ) √n) Also, V( n π§ √π§+π ∑π’=π π π’ π(π§+√π§) √n+2 ∑i=1 Xi ) 2(n+√n) = = βΉ E(∑ Xi ) = np and V(∑ Xi ) = npq √n+2np 2(n+√n) = 4 4(n+√n) 2 1 +2p √n 1 2(1+ ) √n βΆ π as π βΆ ∞ V(∑ni=1 Xi ) = npq (n+√n) 2 =( pq √n+1) βΆ 0 as π βΆ ∞ π§ ππ’ √π§+π ∑ Hence, πΜπ = π(π§+ π’=π is a consistent estimator of p. π§) √ Example 5.3:- Let X1 , … . , Xn be a random sample from π(π, π 2 ). Show that π 2 = 1 n ∑i=1(ππ − πΜ )2 is a consistent estimator of π2. π Sol: - We know that ππ 2 π2 = Μ 2 ∑n i=1(ππ −π ) π2 2 ~ππ−1 ππ 2 π−1 Therefore, πΈ ( π2 ) = π − 1 βΉ πΈ(π 2 ) = ( ππ 2 π π−1 1 ) π 2 = (1 − π) π 2 βΆ π 2 as π βΆ ∞ 1 1 And π ( π2 ) = 2(π − 1) βΉ π(π 2 ) = 2 ( π2 ) π 4 = 2π 4 (π − π2 ) βΆ 0 as π βΆ ∞ Hence, S2 is a consistent estimator of π2. Example 5.3:- Let X1 , … . , Xn be a random sample from π(π, π), π > 0. Find a consistent estimator of µ2. Sol: Let π = πΜ π 2 , then πΈ(π) = πΈ(πΜ π 2 ) = πΈ(πΜ )πΈ(π 2 ) = π. π = π 2 Now, πΈ(π 2 ) = πΈ[(πΜ π 2 )2 ] = πΈ(πΜ 2 )πΈ(π 2 )2 = {π(πΜ ) + πΈ 2 (πΜ )}{π(π 2 ) + πΈ 2 (π 2 )} 2π 2 π 2 2π 1 = {π + π 2 } {π−1 + π 2 } = π 3 {π(π−1) + (π−1) + π + µ} Therefore, 2 2π 1 π(π) = πΈ(π 2 ) − πΈ 2 (π) = π 3 {π(π−1) + (π−1) + π + µ} − π 4 2 2π 1 = π 3 {π(π−1) + (π−1) + π} βΆ 0 as π βΆ ∞ Hence, π = πΜ π 2 is a consistent estimator of µ2. Example 5.3:- Let X1 , … . , Xn be a random sample from π[0, π]. Examine the consistency of the estimators π1 = π(π) , π2 = π(1) + π(π) , π3 = (π + 1)π(1) and π4 = 2πΜ for estimating π. [ISS 2012] Sol: To check this, first we need to find the mean and variance of each Ti’s. Given that π1 , π2 , … . . , ππ ~π(0, π). Transform Yi = Xi Therefore, πΈ[π(π) ] = π+1 and π[π(π) ] = (π+1)2 (π+2) (1) θ then Y1, Y2, Y3………… Yn ~U(0,1) and Y(j) ~Beta(j, n − j + 1) π π ππ π πΈ(π1 ) = πΈ[π(π) ] = πΈ[ππ(π) ] = π+1 = 1+1/π βΆ π as π βΆ ∞ ππ2 π2 π(π1 ) = π[π(π) ] = π[ππ(π) ] = (π+1)2 (π+2) = (π+1)2 (1+2/π) βΆ 0 as π βΆ ∞ Hence, π1 is consistent. (2) π ππ πΈ(π2 ) = πΈ[π(1) + π(π) ] = πΈ[ππ(1) + ππ(π) ] = π+1 + π+1 = π ππ2 π2 π(π2 ) = π[π(1) + π(π) ] = π[ππ(1) + ππ(π) ] = (π+1)(π+2) = (1+1/π)(π+2) βΆ 0 as π βΆ ∞ Hence, π1 is consistent. (3) πΈ(π3 ) = πΈ[(π + 1)π(1) ] = (π + 1)πΈ[ππ(1) ] = (π+1)π π+1 =π (π+1)2 π2 π2 π(π3 ) = π[(π + 1)π(1) ] = (π + 1)2 π[ππ(1) ] = (π+1)2 (π+2) = (π+2) βΆ 0 as π βΆ ∞ Hence, π3 is consistent. (4) π πΈ(π4 ) = πΈ[2πΜ ] = 2πΈ[πΜ ] = 2 2 = π 2 2 4π π π(π4 ) = π[2πΜ ] = 4π[πΜ ] = 12π = 3π βΆ 0 as π βΆ ∞ Hence, π4 is consistent. Example 5.4:- Let X1 , … . , Xn (n > 4) be a random sample from a population with mean µ and variance π2. Consider the following estimators of µ n 1 1 3 1 (π2 +. . . +ππ−1 ) + ππ π = ∑ ππ , W = π1 + π 8 4(n − 2) 8 i=1 (1) Examine whether the estimators U and W are unbiased (2) Examine whether the estimators U and W are consistent (3) Which of these two estimators is more efficient? Sol: - We are given that πΈ(ππ ) = π and π(ππ ) = π 2 1 (1) πΈ(π) = πΈ (π ∑ni=1 ππ ) = π 1 [JAM 2012] ∀ π = 1,2, … and 3 1 1 3 1 πΈ(W) = E [8 π1 + 4(n−2) (π2 +. . . +ππ−1 ) + 8 ππ ] = 8 µ + 4(n−2) (n − 2)µ + 8 µ = µ Hence, U and W both are unbiased estimator of µ 1 (2) Also, π(π) = π (π ∑ni=1 ππ ) = 1 π2 π βΆ 0 as π βΆ ∞ 3 1 and π(W) = V [8 π1 + 4(n−2) (π2 +. . . +ππ−1 ) + 8 ππ ] 1 9 1 = 64 π 2 + 16(n−2)2 (n − 2)π 2 + 64 π 2 9 1 = 16(n−2) π 2 + 32 π 2 18+n−2 = 32(n−2) π 2 16+n = 32(n−2) π 2 β 0 as π βΆ ∞ Hence, U is a consistent estimator but W is not. (3) Suppose, π(π) < π(π) βΉ π2 π 16+n < 32(n−2) π 2 βΉ π2 + 16π > 32π − 64 βΉ π2 − 16π + 64 > 0 βΉ (n − 8)2 > 0 which is true. So U is more efficient than W because π(π) < π(π). Exercise 1.1:- Let π1 , π2 , … . . , ππ are a random sample from geometric distribution π(π = π₯) = ππ π₯ , π₯ = 0,1,2 … …. Find the unbiased estimator of π based on π = ∑ππ=1 ππ . 1.2:- Let π1 , π2 , … . . , ππ are a random sample from geometric distribution π(π = π₯) = ππ π₯−1 , π₯ = 1,2 … …. π Find the unbiased estimator of π based on π = ∑ππ=1 ππ and find its distribution. 1.2:- Let π1 , π2 , … . . , ππ be iid π΅ππ(π). Then show that π(π−π) π(π−1) is an unbiased estimator of π(1 − π) and has smaller variance than the estimator ∑ππ=1 ππ . π π§−π 1.3:- If π1 , π2 , … . . , ππ ~πΈπ₯π(π). Then show that [π − π§πΜ ] Μ > 1 and zero when π§π Μ < 1. when π§π π π (1 − π), where π = π is used to estimate π−π 1.4:- Let π1 , π2 , … . . , ππ are a random sample from the population with probability density function, π(π₯, π) = (π + 1)π₯ π ; 0 < π₯ < 1; π > −1 −(n−1) Show that [ ∑ logx − 1] is an unbiased estimator of π. i 1.5:- Let π1 , π2 , … . . , ππ be a random sample from Laplace distribution, 1 π(π₯, π) = 2 π −|π₯−π| ; −∞ < π₯, π < ∞; Let π(1) ≤. . . . ≤ π(π) be the order statistic. Show that π» = π(1) +π(π) 2 is unbiased for π. 1.6:- Let πΏπ , πΏπ , … . . , πΏπ be a random sample from π΅(π, ππ ), with known mean μ. Find a constant C such that πΆππ is an unbiased estimator of σ, where ππ = π π§ ∑ |πΏπ − µ|. π§ π’=π 2 Hints: πΈ(ππ ) = mean deviation about mean = √π π 1.7 Let πΏπ , πΏπ , πΏπ be a random sample of size 3 from a population with mean μ and variance π2. Let π»π , π»π and π»π are the estimators used to estimate the mean µ, π where π»π = πΏπ + πΏπ − πΏπ π»π = ππΏπ + ππΏπ − ππΏπ π»π = π (ππΏπ + πΏπ + πΏπ ) (a) Are π»π and π»π unbiased estimates? (b) Find the value of π such that π»π is an unbiased estimator for µ. (c) Which is the best estimator? [MHPCS 2007] 1.8 Suppose that πΏπ , πΏπ , … . . , πΏπ are a random sample from the uniform distribution Μ is an unbiased estimator of θ. on (θ − 1, θ + 1). Show that the sample mean πΏ Let πΏ(π) and πΏ(π) be the smallest and largest order statistics derived from πΏπ , … . . , πΏπ . Show also that random variable π΄ = (πΏ(π) + πΏ(π) )/π is an unbiased estimator of θ. Hints: Let π = (π−π+1) 2 ~π(0,1). Then π(π) ~π΅ππ‘π(π, π − π + 1) and 2π(π) + π − 1 = π(π) βΉ π(1) + π(π) = 2(π(1) + π(π) ) + 2π − 2 and πΈ[π(π) ] = π π+1 1.9:- Let πΏπ , πΏπ , … . . , πΏπ be a random sample from a π΅(π, ππ ) distribution, where Μ )π will both µ and ππ are unknown. Then for what value of πΌ, π»πΆ = π ∑π§π’=π(πΏπ − πΏ have minimum mean squared error for estimating ππ , 1.10:- Let π1, π2 , π3 be iid π΅(0, π 2 ) random variables, π>0. Then find the value of k such that the estimator π ∑ππ=π |πΏπ | is an unbiased estimator of π. [JAM 2011] π 2 Hints: Ans is √18. Use the fact that MD about mean for π(µ, π 2 ) is √π π. 1.11:- Let the random variable πΏ~πΌ(5,5 + π). Based on a random sample of size 1, say π1, find an unbiased estimator of π½π is [JAM 2012] Hints: Let Y=X-5, then π~π(0, π). Compute πΈ(π 2 ) = π2 3 . Now adjust. Jam 2005 2.1:- Let π1, π2 ~π΅ππ(π). Show that π = π1, +2π2 is not sufficient for p. 2.2:- Let π1, π2 , π3 ~π΅ππ(π). Examine whether the following are sufficient statistics for p or not? 1 (a) π = 6 (π1, +2π2 + 3π3 ) (b) π = π1 π2 + π3 (c) π = π1 + 2π2 + π3 (d) π = 2π1 + 3π2 + 4π3 Hints: (a) No (b) No 2.3:- Let π1 , … , π4 ~π΅ππ(π). Show that π = π1 (π3 + π4 ) + π2 is not sufficient for p. 2.4:- If π1, π2 ~π(µ, 1). Then show that π = π1 +π2 is sufficient for µ. Hints: Find the conditional distribution of (π1, π2 )|π. 2.5:- Let π1 , π2 , … . . , ππ are iid from π΅ππ(π), π1 , π2 , … . . , ππ are iid from π©ππ(π), and πΏ′π and independent of π′π where π < π < 1 is unknown parameter with π = π − π. π Using the conditional distribution approach show that, π» = ∑π π=π πΏπ − ∑π=π ππ is sufficient for p. Hints: Write T as, π = π1 + β― +ππ + (1 − π1 ) + β― + (1 − ππ ) − π and then show that P(T = t) = (m+n )pn+t (1 − p)m−t . Now use conditional distribution approach to n+t prove this. 2.6:- Suppose π π , π π are iid π(µ, π) , − ∞ < π < ∞. Define π»π = ππ π + π π and ππ = π π + ππ π . Find the conditional distribution of π»π |π»π . Hence comment on π»π . 4 9 Hints: π1 |π2 ~π (3π + (π2 − 3π), ) 5 5 2.7:-If T is sufficient for π, Show that T is also sufficient for πΈ(π). 2.8:- Suppose that X is π΅(π, ππ ) where 0 < π < ∞ is the unknown parameter. By means of the conditional distribution approach, show that | X | is sufficient for π2. Hints: Write the pdf in terms of |X| and use factorization theorem. 2.9:- Let πΏπ , πΏπ , … . . , πΏπ are iid π·ππ(π), ππ , ππ , … . . , ππ are iid π·ππ(ππ), and πΏ′π are independent of π′π where π(> 0) is unknown. Using the conditional distribution π approach show that, π» = ∑π π=π πΏπ + ∑π=π ππ is sufficient for π. Hints: π~π·ππ(ππ + πππ). Use conditional distribution approach to prove this. 2.10:- Let X1 , X2 , … … , Xn be independent random samples with densities Prove that π = π¦π’π§π’ (π π’ /π’) is a sufficient statistics for π. Hints: Use factorization theorem to prove this. 2.11:- Suppose that π π , π π , … … , π π§ are iid with πΌ[π€π, (π€ + π)π] where k is an integer and π(> 0) is the unknown. Find the sufficient statistics for π. Hints: Use factorization theorem to prove π = (X(1) , X(n) ) is sufficient statistics. 2.12:- Suppose that π π , π π , … … , π π§ are iid with πΌ[−π’(π − π), π’(π + π)] where i is an integer and π(> 0) is the unknown. Show that π = (X(1) /i, X(n) /i) is sufficient statistics for π. Hints: Use factorization theorem to prove this. 2.13:- Suppose that π π , π π , … … , π π§ be a random sample from the distribution having the density function 1 π(x; θ) = 2 e−|x−θ| ; −∞ < π₯ < ∞, Show that order statistics, π (π) , π (π) , … … , π (π§) are the only sufficient statistics for π. Hints: Use factorization theorem to prove this. 2.14:- Let π π , π π , … … , π π§ be the random samples from the distribution having pmf π(X = x) = (1 − p)px−θ ; 0 < π < 1, π₯ = π, π + 1, … …, Find the joint sufficient statistics for (π, p). Hints: Use factorization theorem to prove π = (X(1) , ∑ni=1 Xi ) is sufficient. 2.15:- Let π π , π π , … … , π π§ be the random samples from logNormal(µ, π2). Show that π» = (∑ni=1 πππ2 Xi , ∑ni=1 πππXi ) is the joint sufficient statistics for π=(µ, π2). Hints: Use factorization theorem to prove this. 2.16:- Let π π and π π be independent discrete uniform random variable on {π, π, … . . , π} where N is an unknown integer. Obtain the conditional distribution of (π π , π π ) given π = π¦ππ±(π π , π π ). Hence show that T is sufficient statistics for N but not π π + π π . Hints: First show that 2t−1 P(max(X1 , X2 ) = t) = P(X1 = t, X2 < π‘) + P(X1 < π‘, X2 = t) + P(X1 = t, X 2 = t) = N2 . Then show that conditional distribution of (X1 , X2 ) given T is independent of N. 2.17:- Let π π , π π , … … , π π¦ are iid Gamma(πΌ, π½), ππ , ππ , … . . , ππ are iid Gamma(πΌ, kπ½), and πΏ′π are independent of π′π where π < πΌ, π½ < ∞ and π½ is only unknown parameter. Assume that k (>0) is known. Then what is sufficient for π½. Hints: In the joint likelihood of X and Y, apply factorization theorem. 2.18:- Show that both the geometric mean π§√π π , π π , … … , π π§ of the observations and the arithmetic average of the logarithms of the observations are sufficient statistics for π½ when sampling from the Gamma(πΌ, π½) distribution with known πΌ. 1 Hints: π ∑ni=1 πππXi = log[(∏ni=1 Xi )1/n ] and log(.) is a 1-1 function. 2.19:- Let π π , π π , … … , π π§ be iid with common density function is given by 1 π(x; θ) = 2θ e−|x|/θ ; −∞ < π₯ < ∞, 1 1 Define, πΊ = n ∑ni=1 Xi and π = n ∑ni=1 |Xi |. Show that in each case, if it is unbiased estimator and sufficient statistics for π. Hints: T is both unbiased estimator and sufficient whereas S is not. 1 2.20:- Suppose π π has density π1 (x) = θ e−x/θ , x > 0 and π π has density π2 (x) = 2 −2x/θ θ e , x > 0 and π π , π π are independent. Then find a sufficient statistics for π. 2.21:- Suppose π π and π π are iid random variables each following exponential distribution with mean οΏ½. Then which of the following is true? Conditional distribution of π π | π π + π π = π is: π (a) Exponential with mean π and hence π π + π π is sufficient for π. (b) Exponential with mean ππ π and hence π π + π π is not sufficient for π. (c) Uniform(0,t) and hence π π + π π is sufficient for π. (d) Uniform(0,tπ) and hence π π + π π is not sufficient for π. 4.1:- Let X be a random variable having probability mass function 2θ if x = −1 π(π₯) = {θ2 if x = 0 2 1 − 2θ − θ if x = 1 where π ∈ [π, √π − π]. Show that there is only one, and only one, unbiased estimator of (π + π)π based on single observation. [JAM 2009] Hints: 4.2:- Suppose that X1 , X2 , … … , Xn are iid with πΌ(π − π, π + π), where π(> 0) is the unknown parameter. Show that π» = (π (π) , π (π§) ) is jointly sufficient for π but T is not complete. 2π Hints: Use factorization theorem to prove sufficiency. πΈ(X(1) ) = (1 + π) − π+1 and 2π πΈ(X(n) ) = π+1 − (1 − π) βΉ πΈ(X(n) − X(1) ) = 2(π−1) π+1 exactly. 2(π−1) π+1 βΉ πΈ (X(n) − X(1) − 2(π−1) π+1 ) = 0 where as X(n) − X(1) ≠ 4.3:- Let the pdf of X be π(x; θ) = Q(θ)M(x); 0 < π₯ < θ, 0 < θ < ∞, M(x)(> 0) being a continuous function of x. Show that π» = πΏ(π) is complete sufficient statistics for π. Hints: Use the same methodology as we applied for U(0, π). 4.4:- Check whether the following family is complete or not. (a) π(X = 0) = p, π(X = 1) = 3p, π(X = 2) = 1 − 4p; 2 (b) π(X = 0) = p, π(X = 1) = p , π(X = 2) = 1 − p − p2 ; p p (c) π(X = −1) = 2 , π(X = 0) = 1 − p, π(X = 1) = 2 ; |x| θ (d) π(X; θ) = (1 − θ) {2(1−θ)} ; 0 < π < 0.25 0<π<π 0<π<1 −∞ < π₯ < ∞ (e) π π , π π , … … , π π§ are iid π΅(θ, 1), θ = {1,2} 1 1 (f) π~π΅ππ(2, θ), θ = {2 , 4} 1 (g) π~π΅ππ(2, θ), θ = {2 , 1 3 1 , 4} (h) π~π(θ, θ + 1) , − ∞ < π < ∞ (i) π π , π π , … … , π π§ are iid π΅(θ, θ2 ) (j) π π , π π , … … , π π§ are iid π΅(θ, θ2 ) Hints: (a) (c) and (d) (g) are complete, (b), (e), (f), (h) (i) is not. 5.1:- Let that X1 , X2 , … … , Xn be iid random variables with probability density function 1 π(x) = 2 λ3 x 2 e−λx ; π₯ > 0, π > 0 Then which of the following are true? π π (a) π ∑π§π’=π π is an unbiased estimator of π. (b) (c) (d) ππ§ π’ is an unbiased estimator of π. ∑π§ π’=π π π’ π π§ π ∑ π π’=π π π’ ππ§ is ∑π§ π’=π π π’ is a consistent estimator of π. a consistent estimator of π 5.2:- Let that X1 , X2 , … … , Xn (π ≥ 3) be a random sample from U(π-5, π-3) random variables. Let X(1) and X(n) denote the smallest and largest of the sample values. Then which of the following are always true? (a) (X(1) , X(n) ) is complete sufficient for π. (b) X1 + X2 − 2X3 is an ancillary statistics. (c) X(n) + 3 is unbiased for π. (d) X(1) + 5 is consistent for π. 5.3:- If X1 , X2 , … … , Xn are iid from N(0,1). Specify which of the following are unbiased, consistent and sufficient. (a) π»π = X1 +2X2 3 (b) π»π = ∑n i=1 Xi n (c) π»π = ∑n i=1 Xi n+1 [ISS 2010]