STATISTICAL INFERENCE PART II SOME PROPERTIES OF ESTIMATORS 1 SOME PROPERTIES OF ESTIMATORS • θ: a parameter of interest; unknown • Previously, we found good(?) estimator(s) for θ or its function g(θ). • Goal: • Check how good are these estimator(s). Or are they good at all? • If more than one good estimator is available, which one is better? 2 SOME PROPERTIES OF ESTIMATORS • UNBIASED ESTIMATOR (UE): An estimator ˆ is an UE of the unknown parameter , if E ˆ for all Otherwise, it is a Biased Estimator of . Bias ˆ E ˆ ˆ 0. If ˆ is UE of , Bias Bias of ˆ for estimating 3 SOME PROPERTIES OF ESTIMATORS • ASYMPTOTICALLY UNBIASED ESTIMATOR (AUE): An estimator ˆ is an AUE of the unknown parameter , if ˆ 0 Bias ˆ 0 but lim Bias n 4 SOME PROPERTIES OF ESTIMATORS • CONSISTENT ESTIMATOR (CE): An estimator ˆ which converges in probability to an unknown parameter for all is called a CE of . p ˆ . For large n, a CE tends to be closer to the unknown population parameter. • MLEs are generally CEs. 5 EXAMPLE For a r.s. of size n, E X X is an UE of . By WLLN, p X X is a CE of . 6 MEAN SQUARED ERROR (MSE) • The Mean Square Error (MSE) of an estimatorˆ for estimating is MSE ˆ E ˆ Var ˆ Bias ˆ 2 2 If MSE ˆ is smaller, ˆ is a better estimator of . For two estimators, ˆ1 and ˆ2 of , if MSE ˆ1 MSE ˆ2 , ˆ1 is better estimator of . 7 MEAN SQUARED ERROR CONSISTENCY ˆ is called mean squared error consistent (or consistent in quadratic mean) if E{ ˆ }2 0 as n. Theorem: ˆ is consistent in MSE iff i) Var(ˆ )0 as n. ii ) lim E (ˆ) n If E{ ˆ }20 as n, ˆ is also a CE of . 8 EXAMPLES X~Exp(), >0. For a r.s of size n, consider the following estimators of , and discuss their bias and consistency. n ˆ1 X i 1 n n i , ˆ2 X i 1 i n 1 9 SUFFICIENT STATISTICS • X, f(x;), • X1, X2,…,Xn • Y=U(X1, X2,…,Xn ) is a statistic. • A sufficient statistic, Y, is a statistic which contains all the information for the estimation of . 10 SUFFICIENT STATISTICS • Given the value of Y, the sample contains no further information for the estimation of . • Y is a sufficient statistic (ss) for if the conditional distribution h(x1,x2,…,xn|y) does not depend on for every given Y=y. • A ss for is not unique: • If Y is a ss for , then any 1-1 transformation of Y, say Y1=fn(Y) is also a ss for . 11 SUFFICIENT STATISTICS • The conditional distribution of sample rvs given the value of y of Y, is defined as f x1 , x2 , , xn , y; h x1 , x2 , , xn y g y; L ; x1 , x2 , , xn h x1 , x2 , , xn y g y; • If Y is a ss for , then Not depend on for every given y. L ; x1 , x2 , , xn h x1 , x2 , , xn y H x1 , x2 , , xn g y; ss for may include y or constant. 12 • Also, the conditional range of Xi given y not depend on . SUFFICIENT STATISTICS EXAMPLE: X~Ber(p). For a r.s. of size n, n show that Xi is a ss for p. i 1 13 SUFFICIENT STATISTICS • Neyman’s Factorization Theorem: Y is a ss for iff L k1 y; k2 x1 , x2 , The likelihood function Does not depend on xi except through y , xn Not depend on (also in the range of xi.) where k1 and k2 are non-negative functions. 14 EXAMPLES 1. X~Ber(p). For a r.s. of size n, find a ss for p if exists. 15 EXAMPLES 2. X~Beta(θ,2). For a r.s. of size n, find a ss for θ. 16 SUFFICIENT STATISTICS • A ss, that reduces the dimension, may not exist. • Jointly ss (Y1,Y2,…,Yk ) may be needed. Example: Example 10.2.5 in Bain and Engelhardt (page 342 in 2nd edition), X(1) and X(n) are jointly ss for • If the MLE of exists and unique and if a ss for exists, then MLE is a function of a ss for . 17 EXAMPLE X~N(,2). For a r.s. of size n, find jss for and 2. 18 MINIMAL SUFFICIENT STATISTICS • If S( x ) (s1( x ),...,s k ( x )) is a ss for θ, then, ~ ~ ~ S* (x) (s0 (x),s1(x),...,s k (x)) is also a ss ~ ~ ~ ~ for θ. But, the first one does a better job in data reduction. A minimal ss achieves the greatest possible reduction. 19 MINIMAL SUFFICIENT STATISTICS • A ss T(X) is called minimal ss if, for any other ss T’(X), T(x) is a function of T’(x). • THEOREM: Let f(x;) be the pmf or pdf of a sample X1, X2,…,Xn. Suppose there exist a function T(x) such that, for two sample points x1,x2,…,xn and y1,y2,…,yn, the ratio f x1 , x2 , , xn ; f y1 , y2 , , yn ; is constant with respect to iff T(x)=T(y). Then, T(X) is a minimal sufficient statistic for . 20 EXAMPLE • X~N(,2) where 2 is known. For a r.s. of size n, find minimal ss for . Note: A minimal ss is also not unique. Any 1-to-1 function is also a minimal ss. 21 ANCILLARY STATISTIC • A statistic S(X) whose distribution does not depend on the parameter is called an ancillary statistic. • Unlike a ss, an ancillary statistic contains no information about . 22 Example • Example 6.1.8 in Casella & Berger, page 257: Let Xi~Unif(θ,θ+1) for i=1,2,…,n Then, range R=X(n)-X(1) is an ancillary statistic because its pdf does not depend on θ. 23 COMPLETENESS • Let {f(x; ), } be a family of pdfs (or pmfs) and U(x) be an arbitrary function of x not depending on . If E U X 0 for all requires that the function itself equal to 0 for all possible values of x; then we say that this family is a complete family of pdfs (or pmfs). E U X 0 for all U x 0 for all x. i.e., the only unbiased estimator of 0 is 0 itself. 24 EXAMPLES 1. Show that the family {Bin(n=2,); 0<<1} is complete. 25 EXAMPLES 2. X~Uniform(,). Show that the family {f(x;), >0} is not complete. 26 COMPLETE AND SUFFICIENT STATISTICS (css) • Y is a complete and sufficient statistic (css) for if Y is a ss for and the family g y; ; is complete. The pdf of Y. 1) Y is a ss for . 2) u(Y) is an arbitrary function of Y. E(u(Y))=0 for all implies that u(y)=0 27 for all possible Y=y. BASU THEOREM • If T(X) is a complete and minimal sufficient statistic, then T(X) is independent of every ancillary statistic. • Example: X~N(,2). X : the mss for X ~ N( , 2 / n ) and family of N( , 2 / n ) is complete family. X is a complete statistic (n-1)S2/ 2 ~ 2 n 1 S2 Ancillary statistic for By Basu theorem, X and S2 are independent. 28 BASU THEOREM • Example: X1, X2~N(,2), independent, 2 known. • Let T=X1+ X2 and U=X1 - X2 • We know that T is a complete minimal ss. • U~N(0, 22) distribution free of T and U are independent by Basu’s Theorem 29 Problems • Let X1, X 2 ,..., X n be a random sample from a Bernoulli distribution with parameter p. a) Find the maximum likelihood estimator (MLE) of p. b) Is this an unbiased estimator of p? 30 Problems • If Xi are normally distributed random variables with mean μ and variance σ2, what is an unbiased estimator of σ2? 31 Problems • Suppose that X1, X 2 ,..., X n are i.i.d. random variables on the interval [0; 1] with the density function, where α> 0 is a parameter to be estimated from the sample. Find a sufficient statistic for α. 32