Comparison of Improved Estimators under Pitman Closeness Criterion M.DUBE Department of Statistics M.D.University, Rohtak 124001 INDIA S.CHANDRA Department of Mathematics and Statistics Banasthali University, Rajasthan304022 INDIA Abstract The article analyzes the performance of various improved estimators under balanced loss proposed by Zellner. The quantities involved in the balanced loss function have quadratic loss structure, therefore they may fail to reveal some intrinsic properties of estimators. As an alternative it is recommended to use the criterion like the probability of concentration around the true parametric value and closeness of the estimators is the true parametric value, one such criterion proposed by Pitman (1937) called Pitman closeness criterion. The present article derived the asymptotic approximation for the Pitman closeness probability and compared the least squares and Stein rule estimators utilizing the exact prior information binding the regression coefficient using Pitman closeness criterion under balanced loss function using small sigma disturbance asymptotics and disturbance distribution to be not necessarily normal. INTRODUCTION: In regression analysis we are often interested in using an estimator which is precise and which simultaneously provides a model with good fit. Zellner [11] advocated that both the attributes should be utilized in analyzing the performance of estimators. Accordingly, he recommended the use of (quadratic) balanced loss function which maintains a good balance between the two by measuring the precision of estimators and the goodness of the fitted model simultaneously. Many authors used the balanced loss function as a comparison criterion to compare the performance properties of competitive estimators see [7]. The quantities involved in the balanced loss function have quadratic loss structure, therefore they may fail to reveal some intrinsic properties of estimators as pointed out by Rao [3]. As an alternative it is recommended to use the criterion like the probability of concentration around the true parametric value and closeness of the estimators to the true parametric value. Rao [3] suggested to use one such criterion proposed by Pitman called Pitman closeness criterion. Recently, Pitman closeness criterion received attention of many researchers for a comparison of alternative estimators. For instance Shalabh [6] compared the least squares and Stein-rule estimators using Pitman closeness criterion when the disturbances are not necessarily normal. Van Hoa et al [9] derived the conditions for the dominance of the two stage estimators over the least squares and Stein-rule estimators under generalized Pitman closeness criterion. Until now, the extraneous information about the parameters of interest has not been utilized to study the properties of various competitive estimators. However, in most of the situations some prior information, may or may not be exact, is available in the form of linear constraints on parameters. Therefore, using small sigma asymptotic approximations and assuming the disturbances to be not necessarily normal, the present article derived the asymptotic approximation for the Pitman closeness probability and compared the least squares and Stein rule estimators utilizing the exact prior information binding the regression coefficient using Pitman closeness criterion under balanced loss function. Q XX ' and ~ I y XbR ' y XbR 2k X ' X bR n p2 bR' X 'QXbR (2.6) 2. THE MODEL, ESTIMATORS AND LOSS FUNCTIONS: They have obtained the dominance condition of superiority of Stein – type estimators defined above with that of restricted least squares under general quadratic error loss function. Let us consider the following linear regression model, The balanced loss function for the estimation of by an estimator is specified by y X β+ σu (2.1) where y is an n 1 vector of n observations on the response variable, X is a full column rank matrix of n observations on p explanatory variables, is a p 1 column vector of regression coefficient associated with them, is an n 1 vector of disturbances whose elements are independently and identically distributed assuming finiteness of moments up to order four such that E[ut ] 0, E[ut 2 ] 1, E[ut 3 ] r1 , E[ut 4 ] r2 3 , t=1,2,-------------,n and is an unknown positive scalar. The least squares estimator of is ' b X ' X X ' y (2.2) In the presence of some prior information in the form of restrictions on as (2.3) r Rβ where r is a J 1 vector of known elements and R is a J p known prior information design matrix, the least squares estimator of is given by ' ' . bR b X ' X R ' R X ' X R ' r Rb (2.4) The Stein rule based estimators proposed by Srivastava et al [8] and Mittlehammer et al [9] are given by 2k ˆ bR n p2 y' p X y X ' Xb y 'Qy ' where p x I X X ' X X ' (2.5) ' ' L ˆ , 1 ˆ X ' X ˆ y Xˆ y Xˆ (2.7) where is a non stochastic layer lying between 0 and 1. The first component on the right hand side of (2.7) reflects goodness of fitted model while the second component reflects the precision of estimation. The choice of highly depends on the experimenter and objective of the experiment. Another interesting criterion is the Pitman closeness, following Rao et al [4]., the formal definition of the Pitman closeness criterion is : For any two estimators 1 and 2 of , under the loss function(2.7) , the estimator 1 is said to be Pitman closer to the estimator as compared to 2 if PL 2 , L1, 0 1 2 (2.8) 3. Risk Comparison Of Estimators Under Balanced Loss Function: Much of literature is available to study the relative performances of least squares and Stein rule estimators with respect to the risk under quadratic error loss structure using Pitman closeness under the assumption of non normality of disturbances, Employing large sample properties an attempt has been made to study the relative performances of the same in the presence of exact prior information available in the form of linear constraints under balanced loss function. Under the balanced loss function, the risk of b and bR are given by Rb 2 n p 1 p (3.1) RbR Rb 2 J 2 1 (3.2) respectively. Now, for the comparison of the estimators, let us consider the difference of loss functions as (3.3) L LbR , L ˆ , where 1 bR ' X ' X bR (3.4) and 1 ˆ ' X ' X ˆ ' L ˆ , y Xˆ y Xˆ (3.5) bR and are defined in (2.4) and (2.5) respectively. According to the Pitman closeness criterion Stein rule based estimators behaves better-than the restricted least squares if PL 0 exceeds 0.5. For finding an asymptotic approximation of this 1 n probability, let us assume that X ' X to a finite and nonsingular matrix as and consider the following notations ' tends n ' 1 S X 'X n 1 ' ' 2 X QX n W 1 1 n2 (3.6) 1 u 'u 1 1 n n2 b ' X ' X b R L 2 ˆ bR X 'u R ˆ ˆ 'X'X (3.8) utilizing the notations (3.6), (3.8) can be written as L 1 y ' Qu y' p x y 4k 4k y' p x y n p 2 y' Q y n p 2 (3.9) where using (2.1), we can write y ' Px y = P 2 u ' u u ' X ( X ' X 1) X ' u w n or y ' Px y = 2n 1 X 'u n (3.7) Proof : Substitution of (3.4) and (3.5) in (3.3) leads to L bR , y XbR ' y XbR V (3.7) 1 ' ' 1 E[ D] p J 2 X Qe n k n 1 p 2 1 = 1 n n n p2 (3.10) v' sv n 1 1 1 o n n2 (3.11) and and and n D L 4 3R Therefore it can be said that Stein rule based estimator is superior to restricted least squares estimator compared under balanced loss function if E[D>0] Theorem 1: The asymptotic approximation for E[D] to order O(n-1/2 ) is given by 1 1 y ' Qy n 2 1 2 ' X ' Xv o (3.12) 1 2 n n 2 substituting (3.10), (3.11), and (3.12) in (3.9), we get L= 4k 3 n 2 1 1 ' X ' X ' v n v' ( 2 k 2 x' x ' xx) w n 1 ' xx ' v (3.13) E[ D] E[ A] 1 E[ B ] (3.18) n Substitution of values of expectation from (3.16) and (3.17) in (3.18) gives the result (3.7) of theorem 1. From the result (3.7) we observe that ̂ performs better than bR according to the order of our approximation, if Now writing , D= n (3.14) .L 3 4k It is obvious from (3.14) that P[D>0]=P[L>0], Therefore, Stein rule behaves better than restricted least squares in restricted regression if E[D]>0 0 k (1 )[( p J 2) 1 ' X ' Qe] (3.19) for the normal distribution (i.e. 1=0), the condition (3.19) reduces to 0<k<(1-α) (p-J-2) Rewrite D as If no prior information is available regarding the parameter vector (i.e. J=0), the condition reduces to 1 O n n B DA 0<k<(1-α) (p-2) where A= 1 which indicate that the dominance condition for the Stein rule based estimator over least squares estimator shrinks when compared under balanced loss function as compared to that of obtained under quadratic loss structures measuring only the precision of estimation. ' X ' X ' v And B = 1 2 v' 2 XX ' ' X ' X v k 1 n ' X ' Xv.w (3.15) Taking expectation of above as E[A] = 0 E[ B] ; p>2 (3.16) 1 p J 2 1 ' X 'e k (3.17) v 4. GENERAL PITMAN CLOSENESS UNDER THE BALANCED LOSS FUNCTION: Following Rao et al.[4], Stein rule estimator dominates least squares under general Pitman closeness criterion when P[D>0] exceeds 0.5. As it is difficult to derive the exact expressions for probability P[D>0], we consider the approximation when n is large. Theorem 2: The asymptotic expressions for the 1 profability P[D>0] to order O n 2 is given by (4.1) P[D>0] = 0.5 + (1 )( p J 1) 1 m 2 1 k 6 n (4.1) Stein rule estimator performs better restricted least squares when = 6 1 m 1 1 mo 6 n K4 = E[D- E(D)]4 – 3 (E(D)-E(D))2]2 = O(n-1) Employing the above results, the cumulant generating function of D, will be m 0<k< (1 )( p J 1) 1 6 (4.2) K(t) = it j K j j Proof: Suppose g(D) denotes the pdf of D, =i t K1 + therefore P[D>0] = g ( D)dD 0 =(1-α)2 The method of cumulant generating function is adopted to get the probability density function. Before obtaining the cumulants of D, let us first discuss the following results. E[A]=0 E[A2] = (1-α)2 m E[A3] = 1 n 1 p j 2 1 ' X ' Qe k (1 ) p J 4 k 3 1 mo E[A B] = 2 E[A3B] = O (n-1) m= 3 3 t2 ita (it )3 b 2 n 1 ' X ' Qe n a= (1-α) p J 1 r1 1 ' X 'e m mo 1 6 (4.7) t2 ]exp 2 ita (it )3 6 O(n 1 ) n Inverting it to probability density function of D 1 itD (t )dt e 2 along with the results Using above expectations, the cumulants of D 1 1 2 1 itD 2 t dt f ( D) e 2 can be derived upto order O n 2 as K1 = E[D] K2 = E[D-E(D)]2 = (1-α)2 + O(n-1) K3 = E[D- E(D)]3 1 2 1 itD 2 t dt iDf ( D) t.e 2 (4.4) (4.6) The characteristic function of D, upto the order O(n-1/2) can be written as g(D) = 1 ' X ' Qe (I*QXββ’X’Q)e n (4.5) where (t ) exp[ K (t )] =exp[-(1-α)2 E[AB] = O (n-1) where mo = 2 b = (1-α) 1 (4.3) E[B] = it 2 K 2 it 3 K get the It is worth mentioning here that the same conditions of dominance on characterizing scalar is obtained when the estimator defined in (2.6) and bR are compared to the order o n 2 1 2 1 3 itD 2 t dt i 3D D3 f ( D) t .e 2 where f(D) denotes the p.d.f of standard normal variate, we get the following results for g (D) as a b( D3 3) D g(D) = f(D) 1 n 5 o n 2 However, if the terms of (4.8) the difference in dominance conditions may arise. The terms in the risk differences are retained to order o n To the order O(n-1/2 ) P[D>0] = g ( D)dD (4.9) 0 Substitution of the values (4.6) and (4.7) in (4.8) and therefore in (4.9) gives P[D>0]=0.5 + 1 1m (1 )( p J 1) 6 k n 2 are considered 1 2 , resulting no difference in the dominance conditions. At last, on the basis of superiority conditions obtained in (4.11) it can be concluded that if the number of regression coefficients to be estimated are more than one, the least squares estimators can be improved upon by using the Stein rule estimator in the presence of exact prior information binding the regression coefficients using Pitman closeness criterion. which is the result (4.1) of Theorem 2. The condition of superiority of ̂ over bR under balanced loss using Pitman closeness criterion can be obtained as to the order of our approximation as m 0 k (1 )( p J 1) 1 6 1. Mittlehammer, R.C. and Conway, R.K. On the Admissible of Restricted Least Square Estimation, Communications in statistics –Theory and Methods, 13. (1984) 1135-1146. 2. Pitman, E.J.G, The closest estimators of statistical parameters. Proceedings of the Cambridge Philosophical Society, 33 (1937) 212-222. 3. Rao, C.R., Some comments on the minimum MSE as a Criterion of Estimation. In statistics and related topics, eds. N.Esorgo, D.A. Dawson, J.N.K. Rao and A.K. Md, E.Saleh, Amsertdam; North Holland, (1981) 123-143. 4. Rao, C.R., J.P. Keating and R.L. Mason, The Pitman nearness criterion and its determination, communication statistics A – Theory Methods 15(11) (1986) 3173-3191 5. The Stein paradox in the sense of the Pitman Measure of Closeness. The (4.10) provided the quantity on the right hand side is positive. The effect of skewness of the distribution can be seen here. For the symmetrical distribution (i.e. 1 0 ) The condition of dominance reduces to 0<k<(1-α) (p-J-1) REFERENCES (4.11) For the non systematic distribution, if m and 1 have opposite signs i.e m is positive for negatively skewed distributions and is negative for positively skewed distributions, the condition (4.10) is satisfied as long as (4.11) holds true. The dominance conditions obtained in (3.20) and (4.1) when compared, suggest that dominance region is expanded when compared using Pitman closeness criterion. annals of statistics, Vol. 17(3), 13751389. 6. Shalabh, Pitman closeness comparison of least squares and Stein rule estimators in linear regression Models with Non normal Disturbances, American Journal of Mathematical and Management Sciences, Vol 21(182) (2001) 89-100. 7. Shalabh and Chaturvedi, Risk and Pitman closeness properties of feasible generated double K-class stimators in linear regression models with non spherical disturbances under balanced loss function, Journal of muttivariate Analysis, 90 (2004) 229-256. 8. Srivastava, V.K. and Srivastava, A.K., Improved estimation of coefficients in regression models with incomplete prior information , Biometrical Journal, 25, 8 (1983), 775-782. 9. Van Hao, T. and Chaturvedi, A. Performance of 2SHI estimator under the generalized Pitman nearness criterion, Working Paper 99-4, Department of Enomics, University of Wollong ong Australia. (1999). 10. Wan, A.T.K., Risk comparison of the inequality constrained least squares and other related Estimators Under balanced loss , Economics Letters, 46, (1994) 203-210. 11. Zellner, A., Bayesian and non Bayesian using balanced loss functions, in Statistical Decision Theory and Related Topics, V. ecls, S.S. Gupta and J.O. Berger, (Springer- Verlag ), (1994) 377-390.