Discussion Paper: 2011/13 The performance of tests on endogeneity of subsets of explanatory variables scanned by simulation Jan F. Kiviet and Milan Pleus www.ase.uva.nl/uva-econometrics Amsterdam School of Economics Department of Economics & Econometrics Valckenierstraat 65-67 1018 XE AMSTERDAM The Netherlands The performance of tests on endogeneity of subsets of explanatory variables scanned by simulation Jan F. Kiviet and Milan Pleusy 30 December 2014 JEL-code: C01, C12, C15, C30 Keywords: bootstrapping, classi…cation of explanatories, DWH orthogonality tests, test implementation, test performance, simulation design Abstract Tests for classi…cation as endogenous or predetermined of arbitrary subsets of regressors are formulated as signi…cance tests in auxiliary IV regressions and their relationships with various more classic test procedures are examined and critically compared with statements in the literature. Then simulation experiments are designed by solving the data generating process parameters from salient econometric features, namely: degree of simultaneity and multicollinearity of regressors, and individual and joint strength of external instrumental variables. Next, for various test implementations, a wide class of relevant cases is scanned for ‡aws in performance regarding type I and II errors. Substantial size distortions occur, but these can be cured remarkably well through bootstrapping, except when instruments are weak. The power of the subset tests is such that they establish an essential addition to the well-known classic full-set DWH tests in a data based classi…cation of individual explanatory variables. 1. Introduction In this study various test procedures are derived and examined for the classi…cation of arbitrary subsets of explanatory variables as either endogenous or predetermined with respect to a single adequately speci…ed structural equation. Correct classi…cation is Emeritus Professor of Econometrics, Amsterdam School of Economics, University of Amsterdam, Valckenierstraat 65, 1018 XE Amsterdam, The Netherlands (j.f.kiviet@uva.nl) and Visting Professor at the Division of Economics, School of Humanities and Social Sciences, Nanyang Technological University, 14 Nanyang Drive, Singapore 637332 (jfkiviet@ntu.edu.sg). y Amsterdam School of Economics, University of Amsterdam, Valckenierstraat 65, 1018 XE Amsterdam, The Netherlands (m.pleus@uva.nl). Financial support from the Netherlands Organisation for Scienti…c Research (NWO) grant "Statistical inference methods regarding e¤ectivity of endogous policy measures" is gratefully acknowledged. highly important because misclassi…cation leads to either ine¢ cient or inconsistent estimation. The derivations, which in essence are based on employing Hausman’s principle of examining the discrepancy between two alternative estimators, formulate the various tests as joint signi…cance tests of additional regressors in auxiliary IV regressions. Their relationships are demonstrated with particular forms of classic tests such as DurbinWu-Hausman orthogonality tests, Revankar-Hartley covariance tests and Sargan-Hansen overidenti…cation restriction tests. Various di¤erent and some under the null hypothesis asymptotically equivalent implementations follow. The latter vary only regarding degrees of freedom adjustments and the type of disturbance variance estimator employed. We run simulations over a wide class of relevant cases, to …nd out which versions have best control over type I error probabilities and to get an idea of the power of these tests. This should help to use these tests e¤ectively in practice when trying to avoid both evils of inconsistency and ine¢ ciency. To that end a simulation approach is developed by which relevant data generating processes (DGPs) are designed by deriving the values for their parameters from chosen salient features of the system, namely: degree of simultaneity of individual explanatory variables, degree of multicollinearity between explanatory variables, and individual and joint strength of employed external instrumental variables. This allows scanning the relevant parameter space of wide model classes for ‡aws in performance regarding type I and II errors of all implementations of the tests and their bootstrapped versions. We …nd that testing orthogonality by standard methods is impeded for weakly identi…ed regressors. Like bootstrapped tests require resampling under the null, we …nd here that testing for orthogonality by auxiliary regressions bene…ts from estimating variances under the null, as in Lagrange multiplier tests, rather than under the alternative, as in Wald-type tests. However, after proper size correction we …nd that the Wald-type tests exhibit the best power properties. Procedures for testing the orthogonality of all possibly endogenous regressors regarding the error term have been developed by Durbin (1954), Wu (1973), Revankar and Hartley (1973), Revankar (1978) and Hausman (1978). Mutual relationships between these are discussed in Nakamura and Nakamura (1981) and Hausman and Taylor (1981). This test problem has been put into a likelihood framework under normality by Holly (1982) and Smith (1983). Most of the papers just mentioned, and in particular Davidson and MacKinnon (1989, 1990), provide a range of implementations for these tests that can easily be obtained from auxiliary regressions. Although this type of inference problem does address one of the basic fundaments of the econometric analysis of observational data, relatively little evidence on the performance of the available tests in …nite samples is available. Monte Carlo studies on the performance of some of the implementations in static linear models can be found in Wu (1974), Meepagala (1992), Chmelarova and Carter Hill (2010), Jeong and Yoon (2010), Hahn et al.(2011) and Doko Tchatoka (2014), whereas such results for linear dynamic models are presented in Kiviet (1985). The more subtle problem of deriving a test for the orthogonality of subsets of the regressors not involving all of the possibly endogenous regressors has also received substantial attention over the last three decades. Nevertheless, generally accepted rules for best practice on how to approach this problem do not seem available yet, or are confusing as we shall see, and not yet supported by any simulation evidence. Self-evidently, though, the situation where one is convinced of the endogeneity of a few of the regressors, but wants to test some other regressors for orthogonality, is of high practical relevance. 2 If orthogonality is established, this permits to use these regressors as instrumental variables, which (if correct) improves the e¢ ciency and the identi…cation situation, because it makes the analysis less dependent on the availability of external instruments. This is important in particular when available external instruments are weak or of doubtful exogeneity status. Testing the orthogonality of subsets of the possibly endogenous regressors was addressed …rst by Hwang (1980) and next by Spencer and Berk (1981, 1982), Wu (1983), Smith (1984, 1985), Hwang (1985) and Newey (1985), who all suggest various test procedures, some of them asymptotically or even algebraically equivalent. So do Pesaran and Smith (1990), who also provide theoretical arguments regarding an ordering of the power of the various tests, although they are asymptotically equivalent under the null and under local alternatives. Various of the possible sub-set test implementations are paraphrased in Ruud (1984, 2000), Davidson and MacKinnon (1993) and in Baum et al. (2003), and occasionally their relationships with particular forms of Sargan-Hansen (partial-)overidenti…cation test statistics are examined. As we shall show, a few particular situations still call for further analysis and formal proofs, and sometimes results from the studies mentioned above have to be corrected. As far as we know, there are no published simulation results yet on the actual qualities of tests for the exogeneity for arbitrary subsets of the regressors in …nite samples. In this paper we shall try to elucidate the various forms of available test statistics for the endogeneity of subsets of the regressors, demonstrate their origins and their relationships, and also produce solid Monte Carlo results on their performance in single static linear simultaneous models with IID disturbances. That yet no simulation results are available on sub-set tests may be due to the fact that it is not straight-forward how one should design a range of appealing and representative experiments. We believe that in this respect the present study, which closely follows the rules set out in Kiviet (2012), may claim originality. Besides exploiting some invariance properties, we choose the remaining parameter values for the DGP indirectly from the inverse relationships between the DGP parameter values and fundamental orthogonal econometric notions. The latter constitute an insightful base for the relevant nuisance parameter space. The present design can easily be extended to cover cases with a more realistic degree of overidenti…cation and number of jointly dependent regressors. Other obvious extensions would be: to include recently developed tests which are specially built to cope with weak instruments, to consider non Gaussian and non IID disturbances, to examine dynamic models, to include tests for the validity (orthogonality) of instruments which are not included in the regression, etc. Regarding all these aspects the present study just o¤ers an initial reference point. The structure of the paper is as follows. In Section 2, we …rst de…ne the model’s maintained properties and the hypothesis to be tested. Next, in a series of subsections, various routes to develop test procedures are followed and their resulting test statistics are discussed and compared analytically. Section 3 reviews earlier Monte Carlo designs and results regarding orthogonality tests. In Section 4 we set out our approach to obtain DGP parameter values from chosen basic econometric characteristics. A simulation design is obtained to parametrize a synthetic single linear static regression model including two possibly endogenous regressors with an intercept and involving two external instruments. For this design Section 5 presents simulation results for a selection of practically relevant parametrizations. Section 6 produces similar results for bootstrapped versions of the tests, Section 7 provides an empirical case study and Section 8 concludes. 3 2. Testing the orthogonality of subsets of explanatory variables 2.1. The model and setting We consider the single linear simultaneous equation model (2.1) y = X + u; with IID unobserved disturbances u (0; 2 In ); K-element unknown coe¢ cient vector , an n K regressor matrix X and n 1 regressand y: We also have an n L matrix Z containing sample observations on identifying instrumental variables, so E(Z 0 u) = 0; rank(Z) = L; rank(X) = K and rank(Z 0 X) = K: (2.2) In addition, we make asymptotic regularity assumptions to guarantee asymptotic identi…cation of all elements of too and consistency of its IV (or 2SLS) estimator ^ = (X 0 PZ X) 1 X 0 PZ y; (2.3) where PZ = Z(Z 0 Z) 1 Z 0 : Hence, we assume that plim n 1 Z 0 Z = Z0Z and plim n 1 Z 0 X = Z0X (2.4) are …nite and have full column rank, whereas ^ has limiting normal distribution n1=2 ( ^ d ) ! N 0; 2 [ 0 Z0X 1 Z0Z Z0X ] 1 : (2.5) The matrices X and Z may have some (but not all) columns in common and can therefore be partitioned as X = (Y Z1 ) and Z = (Z1 Z2 ); (2.6) where Zj has Lj columns for j = 1; 2: Because the number of columns in Y is K L1 > 0 we …nd from L = L1 + L2 K that L2 > 0; but we allow L1 0; so Z1 may be void. Throughout this paper the model just de…ned establishes the maintained unrestrained hypothesis, which allows Y to contain endogenous variables. Below we will examine particular further curbed versions of the maintained hypothesis and develop tests to verify these further limitations. These are not parametric restraints regarding but involve orthogonality conditions in addition to the L maintained orthogonality conditions embedded in E(Z 0 u) = 0. All these extra orthogonality conditions concern regressors and not further external instrumental variables. Therefore, we consider a partitioning of Y in Ke and Ko columns Y = (Ye Yo ); (2.7) where the variables Ye are maintained as possibly endogenous, whereas for the Ko vari¯ ables Yo their possible orthogonality will be examined, i.e. whether E(Yo0 u) = 0 seems ¯ to hold. We de…ne the n (L + Ko ) matrix Zr = (Z Yo ); 4 (2.8) which relates to all the orthogonality conditions in the restrained model. Note that (2.2) ¯ implies that Zr has full column rank, provided n L + Ko : Now the null and alternative hypotheses that we will examine can be expressed as H 0 : y = X + u; H 1 : y = X + u; u u (0; (0; 2 I); 2 I); E(Zr0 u) = 0; and E(Z 0 u) = 0; E(Yo0 u) 6= 0: (2.9) Hence, H 0 assumes E(Yo0 u) = 0. Under the extended set of orthogonality conditions E(Zr0 u) = 0; i.e. under H 0 ; the restrained IV estimator is ^ = (X 0 PZr X) 1 X 0 PZr y: (2.10) r If H 0 is valid this estimator is consistent and, provided plim n 1 Zr0 Zr = Zr0 Zr exists and is invertible, its limiting normal distribution has variance 2 [ 0Zr0 X Zr01Zr Zr0 X ] 1 ; which involves an asymptotic e¢ ciency gain over (2.5). However, under the alternative hypothesis H 1 estimator ^ r is inconsistent. A test for (2.9) should (as always) have good control over its type I error probability1 and preferably also have high power, in order to prevent the acceptance of an inconsistent estimator. In practice inference on (2.9) usually establishes just one link in a chain of tests to decide on the adequacy of model speci…cation (2.1) and the maintained instruments Z; see for instance Godfrey and Hutton (1994) and Guggenberger (2010). Many of the …rm results obtained below require to make the very strong assumptions embedded in (2.1) and (2.2) and leave it to the practitioner to make a balanced use of them within an actual modelling context. In the derivations to follow we make use of the following three properties of projection matrices, which for any full column rank matrix A are denoted as PA = A(A0 A) 1 A0 : For a full column rank matrix C = (A B) one has (i) PA = PC PA = PA PC ; (ii) PC = PA + PMA B = P(A MA B) ; where MA = I PA ; (iii) for C = (A B); where A = A BD and D an arbitrary matrix of appropriate dimensions, PC = PB + PMB A = PB + PMB A = PC : 2.2. The source of any estimator discrepancy A test based on the Hausman principle focusses on the discrepancy vector ^ ^r = = = = (X 0 PZ X) (X 0 PZ X) (X 0 PZ X) (X 0 PZ X) X 0 PZ y (X 0 PZr X) 1 X 0 PZr y 1 0 X PZ [I X(X 0 PZr X) 1 X 0 PZr ]y 1 (PZ X)0 u^r 1 (PZ Ye PZ Yo Z1 )0 u^r ; 1 (2.11) where u^r = y X ^ r denotes the IV residuals obtained under H 0 : Although testing whether the discrepancy between these two coe¢ cient estimators is signi…cantly di¤erent from zero is not equivalent to testing H 0 ; we will show that in fact all existing test procedures employ the outcome of this discrepancy to infer on the (in)validity of H 0 . 1 An actual type I error probability much larger than the chosen nominal value would more often than intended lead to using an ine¢ cient estimator. A much lower actual type I error than the nominal level would deprive the test from its power hampering the detection of estimator inconsistency. 5 Because (X 0 PZ X) 1 is non-singular ^ ^ r is close to zero if and only if the K (PZ Ye PZ Yo Z1 )0 u^r is. So, we will examine now when its three sub-vectors Ye0 PZ u^r ; Yo0 PZ u^r and Z10 u^r 1 vector (2.12) will jointly be close to zero. Note that due to the identi…cation assumptions both PZ Ye and PZ Yo will have full column rank so cannot be O. For the IV residuals u^r we have X 0 PZr u^r = 0; and since PZr X = (PZr Ye Yo Z1 ); this yields Ye0 PZr u^r = 0; Yo0 u^r = 0 and Z10 u^r = 0: (2.13) Note that the third vector of (2.12) is always zero according to the third equality from (2.13). Using projection matrix property (ii) and the …rst equality of (2.13), we …nd for the …rst vector of (2.12) that Ye0 PZ u^r = Ye0 (PZr PMZ Yo )^ ur = Ye0 PMZ Yo u^r ; so Ye0 PZ u^r = Ye0 MZ Yo (Yo0 MZ Yo ) 1 Yo0 MZ u^r : (2.14) This Ke element vector will be close to zero when the Ko element vector Yo0 MZ u^r is. Due to the occurrence of the Ke Ko matrix Ye0 MZ Yo as a …rst factor in the right-hand side of (2.14), it seems possible that Ye0 PZ u^r may be close to zero too in cases where Yo0 MZ u^r 6= 0; we will return to that possibility below. For the second vector of (2.12) we …nd, upon using the second equality of (2.13), that Yo0 PZ u^r = Yo0 MZ u^r : (2.15) Hence, the second vector of (2.12) will be close to zero if and only if the vector Yo0 MZ u^r is close to zero. From the above it follows that Yo0 MZ u^r being close to zero is both necessary and su¢ cient for the full discrepancy vector (2.11) to be small. Checking whether Yo0 MZ u^r is close to zero corresponds to examining to what degree the variables MZ Yo do obey the orthogonality conditions, while using u^r as a proxy for u; which is asymptotically valid under the extended set of orthogonality conditions. Note that by focussing on MZ Yo the tested variables Yo have been purged from their components spanned by the columns of Z: Since these are maintained to be orthogonal with respect to u; they should better be excluded from the test indeed. Since the inverse matrix in the right-hand side of (2.11) is positive de…nite, the probability limits of ^ and ^ r will be similar if and only if plim n 1 Yo0 MZ u^r = 0: Regarding the power of any discrepancy based test of (2.9) it is now of great interest to examine whether it could happen under H 1 to have plim n 1 Yo0 MZ u^r = 0: For that purpose we specify the reduced form equations Yj = Z j + (u 0 j + Vj ); for j 2 fe; og; (2.16) where j is an L Kj matrix of reduced form parameters, j is a Kj 1 vector that parametrizes the simultaneity and Vj establishes the components of the zero mean reduced form disturbances which are uncorrelated with u and of course with Z: After this 6 further parametrization the hypotheses (2.9) can now be expressed as H 0 : o = 0 and 0 H 1 : o 6= 0: Let (L + Ko ) (L + Ko ) matrix be such that = (Zr0 Zr ) 1 : From Yo0 PZ [In X(X 0 PZr X) 1 X 0 PZr ]u Yo0 PZ [PZr PZr X(X 0 PZr X) 1 X 0 PZr ]u Yo0 PZ Zr [IL+Ko P 0 Zr0 X ] 0 Zr0 u Yo0 MZ u^r = = = (2.17) it follows that plim n 1 Yo0 MZ u^r = 0 if (L + Ko ) 1 vector plim n 1 Zr0 u = 2 (00 0o )0 is in the column space spanned by plim n 1 Zr0 X = Zr0 X : This is obviously the case when 1 vector Zr0 X c; with o = 0: However, it cannot occur for o 6= 0; because (L + Ko ) c a K 1 vector, has its …rst L K elements equal to zero only for c = 0; due to the identi…cation assumptions. This excludes the existence of a vector c 6= 0 yielding 2 0 0 0 (0 o ) when o 6= 0; so under asymptotic identi…cation the discrepancy will Zr0 X c = be nonzero asymptotically when Yo contains an endogenous variable. Cases in which the asymptotic identi…cation assumptions are violated are e = p p Ce = n and/or o = Co = n; where Ce and Co are matrices of appropriate dimensions with full column rank and all elements …xed and …nite.2 Examining Zr0 X c closer yields Zr0 X c = 0 o Z0Z Z0Z e + e Vo0 Ve + 2 0 o e c1 + Z0Z o Yo0 Yo c2 + 0 o Z 0 Z1 Z 0 Z1 c3 ; (2.18) p where c = (c01 c02 c03 )0 and Vo0 Ve = plim n 1 Vo0 Ve : If only o = Co = n; so when all the instruments Z are weak and asymptotically irrelevant for the set of regressors Yo whose orthogonality is tested, we can set c1 = 0 and c3 = 0 and then for c2 = 2 Yo10 Yo o = 2 ( 2 o 0o + Vo0 Vo ) 1 o 6= 0 we have Zr0 X c = 2 (00 0op )0 6= 0; demonstrating that the test will have no asymptotic power. If only e = Ce = n; thus all the instruments Z are weak for Ye ; a solution c 6= 0 can be found upon taking c2 = 0; c3 = 0 and c1 6= 0, provided Vo0 Ve + 2 o 0e 6= O or Ye and Yo are asymptotically not uncorrelated. Only c3 has to be set at zero to …nd a solution when Z is weak for both Yo and Ye : From (2.18) it can also be established that when from Z2 at least Ke + Ko instruments are not weak for Y the discrepancy will always be di¤erent from zero asymptotically when o 6= 0: Using (2.16) we also …nd plim n 1 Ye0 MZ Yo = 0Vo0 Ve + 2 e 0o , which demonstrates that the …rst vector of (2.12) would for o 6= 0 tend to zero also when e = 0 while the reduced form disturbances of Ye and Yo are uncorrelated. This indicates the plausible result that a discrepancy based test may loose power when Ye is unnecessarily treated as endogenous and Yo is establishing a weak instrument for Ye after partialing out Z. 2.3. Testing based on the source of any discrepancy Next we examine the implementation of testing closeness to zero of Yo0 MZ u^r in an auxiliary regression. Consider y = X + PZ Y o + u ; (2.19) where u = u PZ Yo : Its estimation by IV employing the instruments Zr yields coe¢ cients that can be obtained by applying OLS to the second-stage regression of y on PZr X and PZr PZ Yo = PZ Yo : For partitioned regression yields ^ = (Y 0 PZ MP o 2 Zr X PZ Yo ) 1 Yo0 PZ MPZr X y; (2.20) Doko Tchatoka (2014) considers a similar situation for the special case Ke = 0 and Ko = 1: 7 where, using rule (i), Yo0 PZ MPZr X y = Yo0 PZ [I X(X 0 PZr X) 1 X 0 PZr ]y = Yo0 PZ u^r : Thus, by testing = 0 in (2.19) we in fact examine whether Yo0 PZ u^r = Yo0 MZ u^r di¤ers signi…cantly from a zero vector, which is indeed what we aim for.3 Alternatively, consider the auxiliary regression (2.21) y = X + MZ Yo + v ; where v = u MZ Yo : Using the instruments Zr involves here applying OLS to the second-stage regression of y on PZr X and PZr MZ Yo = PZr Yo PZr PZ Yo = Yo PZ Yo = MZ Yo : This yields ^ = (Y 0 MZ MP X MZ Yo ) 1 Y 0 MZ MP X y; (2.22) o o Zr Zr where Yo0 MZ MPZr X y = Yo0 MZ [I PZr X(X 0 PZr X) 1 X 0 PZr ]y = Yo0 [I X(X 0 PZr X) 1 X 0 PZr ]y Yo0 PZ [I = Yo0 MZ u^r : X(X 0 PZr X) 1 X 0 PZr ]y (2.23) Thus, like testing = 0 in (2.19), testing = 0 in auxiliary regression (2.21) examines the magnitude of Yo0 MZ u^r : The estimator for resulting from (2.21) is ^ = (X 0 PZr MM Yo PZr X) 1 X 0 PZr MM Yo y: r Z Z Because PZr MMZ Yo = PZr PZr PMZ Yo = PZr (PZ + PMZ Yo )PMZ Yo = PZr PMZ Yo = PZ ; we …nd ^ r = ^ : Hence, the IV estimator of exploiting the extended set of instruments in the auxiliary model (2.21) equals the unrestrained IV estimator ^ : Many text books mention this result for the special case Ke = 0: From the above we …nd that testing whether included possibly endogenous variables Yo can actually be used e¤ectively as valid extra instruments, can be done as follows: Add them to Z; so use Zr as instruments, and add at the same time the regressors MZ Yo (the reduced form residuals of the alleged endogenous variables Yo in the maintained model) to the model, and then test their joint signi…cance. When testing = 0 in (2.21) by a Wald-type statistic, and assuming for the moment that 2 is known, the test statistic is 2 0 y PMPZ X MZ Yo y = 2 y 0 (MA MC )y; (2.24) r where A = PZr X; B = MZ Yo and C = (A B): Hence, y 0 PMPZ X MZ Yo y is equal to the r di¤erence between the OLS residual sums of squares of the restricted (by = 0) and the unrestricted second stage regressions (2.21). One easily …nds that testing = 0 in (2.19) by a Wald-type test yields in the numerator y 0 PMPZ r X P Z Yo y = y 0 (MA MC )y; with again A = PZr X = (PZr Ye Yo Z1 ); but C = (A B ) with B = PZ Yo : Although C 6= C, both span the same sub-space, so MC = MC and thus the two auxiliary regressions lead to numerically equivalent Wald-type test statistics. 3 This procedure provides the explicit solution to the exercise posed in Davidson and MacKinnon (1993, p.242). 8 Of course, 2 is in fact unknown and should be replaced by an estimator that is consistent under the null. There are various options for this. Two rather obvious choices would be ^ 2 = u^0 u^=n or ^ 2r = u^0r u^r =n; giving rise to two under the null (and also under local alternatives) asymptotically equivalent test statistics, both with 2 (Ko ) asymptotic null distribution. Further asymptotically equivalent variants can be obtained by employing a degrees of freedom correction in the estimation of 2 and/or by dividing the test statistic by Ko and then confronting it with critical values from an F distribution with Ko and n l degrees of freedom with l some …nite number, possibly K + Ko : Testing the orthogonality of Yo and u; while maintaining the endogeneity of Ye ; by a simple 2 -form statistic and using as in a Wald-type test the estimate ^ 2 (without any degrees of freedom correction) from the unrestrained model, will be indicated by Wo : When using the uncorrected restrained estimator ^ 2r ; the statistic will be denoted here as Do : So we have the two archetype test statistics Wo = y 0 PMPZ r X M Z Yo y=^ 2 and Do = y 0 PMPZ r X M Z Yo y=^ 2r : (2.25) Using the restrained 2 estimator, as in a Lagrange-multiplier-type test under normality, was already suggested in Durbin (1954, p.27), where Ke = L1 = 0 and Ko = L2 = 1: Before we discuss further options for estimating 2 in general sub-set tests, we shall …rst focus on the special case Ke = 0; where the full set of endogenous regressors is tested. Then ^ 2r = y 0 MX y=n = n nK s2 stems from OLS. Wu (1973) suggested for this case four test statistics, indicated as T1 ; :::; T4 ; where T4 = n 2Ko n n L1 1 Do and T3 = Ko 2Ko n L1 1 Wo : Ko (2.26) On the basis of his simulation results Wu recommended to use the monotonic transformation of T4 (or Do ) T2 = T4 1 Ko T n 2Ko L1 4 = n 2Ko n L1 1 Do : Ko 1 Do =n (2.27) He showed that under normality of both structural and reduced form disturbances the null distribution of T2 is F (Ko ; n 2Ko L1 ) in …nite samples.4 Because Ke = 0 implies MPZr X = MX we …nd from (2.24) that in this case Do y 0 PMX MZ Yo y y 0 P M X M Z Yo y y 0 P M X M Z Yo y =n 0 =n 0 = : 1 Do =n y (MX PMX MZ Yo )y y M(X MZ Yo ) y •2 Hence, from the …nal expression we see that T2 estimates 2 by • 2 = y 0 M(X MZ Yo ) y=n; which is the OLS residual variance of auxiliary regression (2.21). Like ^ 2 and ^ 2r ; • 2 is consistent under the null, because plim n 1 Yo0 MZ u^r = 0 implies, after substituting (2.23) in (2.22), that plim ^ = 0: Pesaran and Smith (1990) show that under the alternative plim ^ 2 plim ^ 2r 4 plim • 2 Wu’s T1 test for case Ke = 0, which under normality has a F (Ko ; L2 reputation in terms of power and therefore we leave it aside. 9 Ko ) distribution, has a poor and then invoke arguments due to Bahadur to expect that T2 (which uses • 2 ) has better power than T4 (which uses ^ 2r ), whereas both T2 and T4 are expected to outperform T3 (which uses ^ 2 ). However, they did not verify this experimentally. Moreover, because T2 is a simple monotonic transformation of T4 when Ke = 0; we also know that after a fully successful size correction both should have equivalent power. Following the same lines of thought for cases where Ke > 0; we expect (after proper size correction) Do to do better than Wo ; but Pesaran and Smith (1990) suggest that an even better result may be expected from formally testing = 0 in the auxiliary regression (2.21) while exploiting instruments Zr : This amounts to the 2 (Ko ) test statistic To , which (omitting its degrees of freedom correction) generalizes Wu’s T2 for cases where Ke 0; and is given by To = y 0 PMPZ r X M Z Yo y=• 2 = y 0 (MA MC )y=• 2 ; (2.28) MZ Yo ^)=n: (2.29) with • 2 = (y X^ MZ Yo ^)0 (y X^ Actually, it seems that Pesaran and Smith (1990, p.49) employ a slightly di¤erent estimator for 2 ; namely (y X^ MZ Yo ^ )0 (y X^ MZ Yo ^ )=n (2.30) with ^ = (Y 0 MZ Yo ) 1 Y 0 MZ (y o o X ^ ): (2.31) However, because OLS residuals are orthogonal to the regressors we have Yo0 MZ (y X ^ MZ Yo ^) = 0; from which it follows that ^ = ^ ; so their test is equivalent with To : When Ke > 0 the three tests Wo ; Do and To are not simple monotonic transformations of each other, so they may have genuinely di¤erent size and power properties in …nite samples. In particular, we …nd that for y 0 PC y y 0 PA y Do = 0 ; 1 Do =n (^ ur u^r y 0 PC y + y 0 PA y)=n the denominator in the right-hand expression di¤ers from • 2 (unless Ke = 0):5 Using that ^ is given by (2.31) we …nd from (2.29) that • 2 = u^0 MMZ Yo u^=n ^ 2 ; so Wo To ; (2.32) whereas Do can be at either side of Wo and To : 5 Therefore, the test statistic (54) suggested in Baum et al. (2003, p.26), although asymptotically equivalent to the tests suggested here, is built on an inappropriate analogy with the Ke = 0 case. Moreover, in their formulas (53) and (54) Q should be the di¤erence between the residual sums of squares of second-stage regressions, precisely as in (2.25). The line below (54) suggests that Q is a di¤erence between squared IV residuals (which would mean that Q could be negative) of the (un)restricted auxiliary regressions, although their footnote 25 seems to suggest otherwise. 10 2.4. Testing based on the discrepancy as such Direct application of the Hausman (1978) principle yields the test statistic Ho = ( ^ ^ )0 [^ 2 (X 0 PZ X) r 1 ^ 2r (X 0 PZr X) 1 ] ( ^ ^ ); r (2.33) which uses a generalized inverse for the matrix in square brackets. When 2 were known the matrix in square brackets would certainly be singular though semi-positive de…nite. Using two di¤erent 2 estimates might lead to nonsingularity but could yield negative test statistics. As is obvious from the above, (2.33) will not converge to a 2K distribution under H 0 ; but in our framework to one with Ko degrees of freedom, cf. Hausman and Taylor (1981). Some further analysis leads to the following. Let have separate components as follows from the decompositions X = Ye e + Yo o + Z1 1 =Y eo + Z1 whereas (X 0 PZ X) 1 has blocks Ajk ; j; k = 1; 2; where A11 is a Keo Keo = Ke + Ko : Then we …nd from (2.11) and (2.13) that ^ ^ ^ = (X 0 PZ X) r eo ^ Y 0 PZ u^r 0 1 A11 A21 = (2.34) 1; Keo matrix with Y 0 PZ u^r ; (2.35) = A11 Y 0 PZ u^r : eo;r Hence, the discrepancy vector of the two coe¢ cient estimates of just the regressors in Y; but also those of the full regressor matrix X; are both linear transformations of rank Keo of the vector Y 0 PZ u^r : Therefore it is obvious that the Hausman-type test statistic (2.33) can also be obtained from Ho = ( ^ eo ^ 0 2 0 eo;r ) [^ (Y PMZ1 Z2 Y ) 1 ^ 2r (Y 0 PMZ1 (Z2 Yo ) Y ) 1 ] ( ^ eo ^ eo;r ): (2.36) Both test statistics are algebraically equivalent, because of the unique linear relationship ^ ^ = r IKeo A21 A111 ( ^ eo ^ eo;r ): (2.37) Calculating (2.36) instead of (2.33) just mitigates the numerical problems. One now wonders whether an equivalent Hausman-type test can be calculated on the basis of the discrepancy between the estimated coe¢ cients for just the regressors Yo : This is not the case, because a relationship of the form ( ^ eo ^ eo;r ) = G( ^ o ^ o;r ); where G is a Keo Ko matrix, cannot be found6 . However, a matrix G can be found such that ( ^ eo ^ eo;r ) = G^; indicating that test Ho can be made equivalent to the three distinct tests of the foregoing subsection, provided similar 2 estimates are being used. Using (2.14) and (2.15) in (2.35) we obtain ^ eo ^ eo;r = A11 Y 0 PZ u^r = A11 Ye0 MZ Yo (Yo0 MZ Yo ) IKo 1 (Yo0 MZ MPZr X MZ Yo )^; (2.38) 6 Note that Wu (1983) and Hwang (1985) start o¤ by analyzing a test based on the descripancy ^ ^ : Both Wu (1983) and Ruud (1984, p.236) wrongly suggest equivalence of such a test with o o;r (2.33) and (2.36). 11 because (2.22) and (2.23) yield Yo0 MZ u^r = (Yo0 MZ MPZr X MZ Yo )^: So, under the null hypothesis particular implementations of Wo ; Do ; To and Ho are equivalent.7 When Ho is used with two di¤erent 2 estimates it may come close to a hybrid implementation of Wo and Do where the two residual sums of squares in the numerator are scaled by di¤erent 2 estimates as in W Do = y 0 M(PZr X MZ Yo ) y : ^2 y 0 MPZr X y ^ 2r (2.39) 2.5. Testing based on covariance of structural and reduced form disturbances Auxiliary regression (2.21) is used to detect correlation of u and Vo (the reduced form disturbances of Yo ) by examing the covariance of the residuals u^r and MZ Yo : This might perhaps be done in a more direct way by augmenting regression (2.1) by the actual reduced form disturbances, giving y = X + (Yo where w = u (Yo Z (2.40) can be written as y = Ye = X o) e with a Ko Z o) (2.40) +w ; 1 vector. Let Z + Yo ( o + ) + Z1 ( + Z2 + w 1 o1 ) Z2 o = Z1 o2 o1 + Z2 o2 ; then +w (2.41) in which we may assume that E(Z 0 w ) = 0; though E(Ye0 w ) 6= 0: However, testing = 0; which corresponds to = 0 in (2.40), through estimating (2.41) consistently is not an option, unless Ke = 0. For Ke > 0; which is the case of our primary interest here, (2.41) contains all available instruments as regressors, so we cannot instrument Ye : For the case Ke = 0 the test of = 0 yields the test of Revankar and Hartley (1973), which is an exact test under normality. When Ko = L2 (just identi…cation) it specializes to Wu’s T2 .8 When L2 > Ko (overidenti…cation) Revankar (1978) argues that testing the Ko restrictions = 0 by testing the L2 restrictions = 0 is ine¢ cient. He then suggests to test = 0 by a quadratic form in the di¤erence of the least-squares estimator of o + in (2.41) and the IV estimator of o :9 From the above we see that the tests on the covariance of disturbances do not have a straight-forward generalization for the case Ke > 0: However, a test that comes close to it replaces the L L1 columns of Z2 in (2.41) by a set of L K regressors Z2 which span a subspace of Z2 ; such that (PZ Ye Z1 Z2 ) spans the same space as Z: Testing these 7 This generalizes the equivalence result mentioned below (22.27) in Ruud (2000, p.581), which just treats the case Ke = 0: Note, however, that because Ruud starts o¤ from the full discrepancy vector, the transformation he presents is in fact singular and therefore the inverse function mentioned in his footnote 24 is non-unique (the zero matrix may be replaced with any other matrix of the same dimensions). To obtain a unique inverse transformation, one should start o¤ from the coe¢ cient discrepancy for just the regressors Y; and this is found to be nonsingular for Ke = 0 only. 8 This is proved as follows: Both tests have regressors X under the null, and under the alternative the full column rank matrices (X PZ Yo ) and (X Z2 ) respectively. These matrices span the same space when X = (Yo Z1 ) and Z = (Z1 Z2 ) have the same number of columns. 9 Meepagala (1992) produces numerical results indicating that the descripancy based tests have lower power than the Revankar and Hartley (1973) test when instruments are weak and than the Revankar (1978) test when the instruments are strong. 12 L K exclusion restrictions yields the familiar Sargan-Hansen test for testing all the so-called overidenti…cation restrictions of model (2.1). It is obvious that this test will have power for alternatives in which Z2 and u are correlated, possibly because some of the variables in Z2 are actually omitted regressors. In practical situations this type of test, and also Hausman type tests for the orthogonality of particular instruments not included as regressors in the speci…cation10 , are very useful. However, we do not consider such implementations here, because right from the beginning we have chosen a context in which all instruments Z are assumed to be uncorrelated with u: This allows focus on tests serving only the second part of the two-part testing procedure as exposed by Godfrey and Hutton (1994), who also highlight the asymptotic independence of these two parts. 2.6. Testing by an incremental Sargan test The original test of overidentifying restrictions initiated by Sargan (1958) does not enable to infer directly on the orthogonality of individual instrumental variables, but a so-called incremental Sargan test does. It builds on the maintained hypothesis E(Z 0 u) = 0 and can test the orthogonality of additional potential instrumental variables. Choosing for these the included regressors Yo yields a test statistic for the hypotheses (2.9) which is given by u^0 PZ u^r u^0 PZ u^ : (2.42) So = r 2r ^r ^2 When using for both separate Sargan statistics the same PZ u^ = (PZ PPZ X )y; the numerator would be u^0r PZr u^r 2 estimate, and employing u^0 PZ u^ = y 0 (PZr PPZr X PZ + PPZ X )y = y 0 (PMZ Yo + PPZ X PPZr X )y = y 0 (P(PZ X MZ Yo ) PPZr X )y; whereas that of Wo and Do in (2.24) is given by y 0 (PC PA )y; where C = (A B) 11 with A = PZr X and B = MZ Yo : Equivalence is proved by using general result (iii) on projection matrices, upon taking A = PZ X: Using PZr = PZ + PMZ Yo ; we have A = A PB X = A B(B 0 B) 1 B 0 X; so D = (B 0 B) 1 B 0 X: Thus P(A B) = P(A B) = P(PZ X MZ Yo ) giving u^0r PZr u^r u^0 PZ u^ = y 0 (P(PZr X MZ Yo ) PPZr X )y: (2.43) Hence, in addition to the Ho statistic, So establishes yet another hybrid form combining elements of both Wo and Do ; but di¤erent from (2.39). 2.7. Concerns for practitioners The foregoing subsections demonstrate that all available archetypical statistics Wo ; Do ; To ; Ho and So for testing the orthogonality of a subset of the potentially endogenous 10 See Hahn et al. (2011) for a study on its behaviour under weak instruments. Ruud (2000, p.582) proves this just for the special case Ke = 0: Newey (1985, p.238), Baum et al. (2003, p.23 and formula 55) and Hayashi (2002) mention equivalence for Ke 0, but do not provide a proof. 11 13 regressors basically just di¤er regarding the way in which the expresssions they are based on are scaled with respect to 2 : Both So and Ho (and of course W Do ) show a hybrid nature in this respect, because their most natural implementations require two di¤erent 2 estimates, which may lead to negative test outcomes. In addition to that, Ho has the drawback that it involves a generalized inverse, whereas calculation of the other four is rather straight-forward.12 Similar di¤erences and correspondences carry over to more general models, which would require GMM estimation, see Newey (1985) and Ahn (1997). Although of no concern asymptotically, these di¤erences may have major consequences in …nite samples, thus practitioners are in need of clues which implementations should be preferred.13 Therefore, in the remainder of this study, we will examine the performance in …nite samples of all these …ve archetypical tests. First, we will examine whether any simple degrees of freedom corrections seem to lead to acceptable size control. Next, only for those variants that pass this test we will perform some power calculations. 3. Earlier Monte Carlo designs and results In the literature the actual rejection frequencies of tests on the independence between regressors and disturbances have been examined by simulation only for situations where all possibly endogenous regressors are tested jointly, hence Ke = 0. To our knowledge, sub-set orthogonalty tests have not been examined yet. Wu (1974) was the …rst to design a simulation study in which he examined the four tests suggested in Wu (1973). He made substantial e¤orts, both analytically and experimentally, to assess the parameters and model characteristics which actually determine the distribution of the test statistics and their power curves. His focus is on the case where there is one possibly endogenous regressor (Ko = 1), an intercept and one other included exogenous regressor (L1 = 2) and two external instruments (L2 = 2), giving a degree of overidenti…cation of 1. All disturbances are assumed normal, all exogenous regressors are mutually orthogonal and all consist of elements equal to either 1, 0, or -1, whereas all instruments have coe¢ cient 1 in the reduced form. Wu demonstrates that all considered test statistics are functions of statistics that follow Wishart distributions which are invariant with respect to the values of the structural coe¢ cients of the equation of interest. The e¤ects of changing the degree of simultaneity and of changing the joint strength of the external instruments are examined. Because the design is rather in‡exible regarding varying the explanatory part of the reduced form, no separate attention is paid to the e¤ects of multicollinearity of the regressors on the rejection 12 It is not obvious why Pesaran and Smith (1990, p.49,55) mention that they …nd To a computationally more attractive statistic than Wo : All three test statistics are very easy to compute. However, To is the only one that strictly applies a standard procedure (Wald) to testing zero restrictions in an auxiliary regression, which eases its use by standard software packages. On the other hand Baum et al. (2003, p.26) characterize tests like To as "computationally expensive and practically cumbersome", which we …nd far fetched too. 13 Under the heading of "Regressor Endogeneity test" EViews 8.1 presents statistic So where for both 2 estimates n K degrees of freedom are used, like it does for the J statistic. In Stata 13 the "hausman" command calculates Ho by default and o¤ers the possibility to calculate Wo and Do . The degrees of freedom reported is the rank of the estimated variance of the discrepancy vector. In case of Ho this is not correct. It is possible to overwrite the degrees of freedom by an additional command. The popular package "ivreg2" only reports Do with the correct degrees of freedom. 14 probabilities, nor to the e¤ects of weakness of individual instruments. Although none of the tests examined is found to be superior under all circumstances, test T2 ; which is exact under normality and generalized as To in (2.28), is found to be the preferred one. Its power increases with the absolute value of the degree of simultaneity, with the joint strength of the instruments and with the sample size. Nakamura and Nakamura (1985) examine a design where Ke = 0; Ko = 1; L1 = 2; L2 = 3 and all instruments are mutually independent standard normal. The structural equation disturbances u and the reduced form disturbances v are IID normal with variances 2u and 2v respectively and correlation . They focus on the case where all coe¢ cients in the structural equation and in the reduced form equation for the possibly endogenous regressor are unity. Given the …xed parameters the distribution of the test statistic T2 now depends only on the values of 2 ; 2u and 2v : Attention is drawn to the fact that the power of an endogeneity test and its interpretation di¤ers depending on whether the test is used to signal: (a) the degree of simultaneity expressed as , (b) the simultaneity expressed as the covariance = u v , or (c) the extent of the asymptotic bias of OLS (which in their design is also determined just by ; 2u and 2v ). When testing (a) a natural choice of the nuisance parameters (which are kept …xed when is varied to obtain a power curve) are u and v : However, when testing (b) or (c) ; u and v cannot all be chosen independently. The study shows that, although the power of test T2 does increase for increasing values of 2 while keeping u and v constant, it may decrease for increasing asymptotic OLS bias. Therefore, test T2 is not very suitable for signaling the magnitude of OLS bias. In this design 2v = 5(1 R2 )=R2 ; where R2 is the population coe¢ cient of determination of the reduced form equation for the possibly endogenous regressor. The joint strength of the instruments is a simple function of R2 and hence of v : Again, due to the …xed values of the reduced form coe¢ cients the e¤ects of weakness of individual instruments or of multicollinearity cannot be examined from this design. The study by Kiviet (1985) demonstrates that in models with a lagged dependent explanatory variable the actual type I error probability of test T2 may deviate substantially from the chosen nominal level. Then high rejection frequencies under the alternative have little or no meaning.14 In the present study we will stick to static cross-section type models. Thurman (1986) performs a small scale Monte Carlo simulation of just 100 replications on a speci…c two equation simultaneous model using empirical data for the exogenous variables from which he concludes that Wu-Hausman tests may have substantial power under particular parametrizations and none under others. Chmelarova and Hill (2010) focus on pre-test estimation based on test T2 (for Ko = 1; L1 = 2; L2 = 1) and two other forms of contrast based tests which use an improper number of degrees of freedom15 . Their Monte Carlo design is very restricted, because the possibly endogenous regressor and the exogenous regressor (next to the constant) are uncorrelated, so multicollinearity does not occur, which makes the DGP unrealistic. Moreover, all coe¢ cients in the equation of interest are kept …xed and are such that the 14 Because we could not replicate some of the presented …gures for the case of strong instruments, we plan to re-address the analysis of DWH type tests in dynamic models in future work. 15 This may occur when standard software is employed based on a naive implementation of the Hausman test. Practitioners should be adviced never to use these standard options but always perform tests based on estimator contrasts by running the relevant auxiliary regression. 15 signal to noise ratio is always 1. Therefore, the inconsistency of OLS is relatively large (and in fact equal to the simultaneity correlation coe¢ cient ). Because the sample size is not varied and neither is the instrument strength parameter16 the results do not allow to form an opinion on how e¤ective the T2 test is to diagnose simultaneity. Jeong and Yoon (2010) present a study in which they examine by simulation what the rejection probability of the Hausman test is when an instrument is employed which is actually correlated with the disturbances. Also for the sub-set tests to be examined here the situation seems of great practical relevance that they might be implemented while using some variable(s) as instruments which are in fact endogenous. In our Monte Carlo experiments we will cover such situations, but we do not …nd the design as used by Jeong and Yoon, in which the endogeneity/exogeneity status of variables depends on the sample size very useful. 4. A more comprehensive Monte Carlo design To examine the di¤erences between the various sub-set tests regarding their type I and II error probabilities in …nite samples we want to lay out a Monte Carlo design which is less restrictive than those just reviewed. It should allow to represent the major characteristics of data series and their relationships as faced in empirical work, while avoiding the imposition of awkward restrictions on the nuisance parameter space. Instead of picking particular values for the coe¢ cients and further parameters in a simple DGP, and check whether or not this leads to covering empirically relevant cases, we choose to approach this design problem from the opposite direction. 4.1. The simulated data generating process Model (2.1) is specialized in our simulations to y= y (2) y (3) = = 1 21 31 + + + 2y (2) + 22 z (2) 32 z (2) 3y + + (3) (4.1) + u, 23 z (3) 33 z (3) (2) (4.2) (3) (4.3) +v , +v , where is an n 1 vector consisting of ones. So, K = 3; L1 = 1 and L2 = 2; with Ko + Ke = 2; Y = (y (2) y (3) ); Z1 = and Z = ( z (2) z (3) ): Since K = L; at this stage we only investigate the case in which under the unrestrained alternative hypothesis the single simultaneous equation (4.1) is just identi…ed according to the order condition. Because the statistics to be analyzed will be invariant regarding the values of the intercepts, these are all set equal to zero, thus 1 = 21 = 31 = 0. Ful…llment of the rank condition for identi…cation then implies that the inequality 22 33 6= 23 32 (4.4) has to be satis…ed. The vectors z (2) and z (3) will be generated as mutually independent IID(0; 1) series. They have been drawn only once and then were kept …xed over all replications. In fact 16 If the e¤ects of a weaker instrument had been checked the simulation estimates of the moments of IV (which do not exist, because the model is just identi…ed) would have gone astray. 16 we drew two arbitrary series and next rescaled them such that their sample mean and variance, and also their sample covariance correspond to the population values 0, 1 and 0 respectively. To allow for simultaneity of both y (2) and y (3) ; as well as for any value of the correlation between the reduced form disturbances v (2) and v (3) ; these disturbances have components v (2) = (2) + 2 u and v (3) = (3) + (2) + 3 u, (4.5) (2) (3) where the series ui , i and i will be generated as mutually independent zero mean IID series (for i = 1; :::; n). Without loss of generality, we may choose 2u = 1: Scaling the variances of the potentially endogenous regressors simpli…es the model even further, again without loss of generality. This scaling is innocuous, because it can be compensated by the chosen values for 2 and 3 : We will realize 2y(2) = 2y(3) = 1 by choosing appropriate values for 2(2) > 0 and 2(3) > 0 as follows. For the variance of the IID series for the reduced form disturbances and for the possibly endogenous explanatory variables we …nd 2 v (2) = 2 = 2 (2) 2 y (2) + 2 2; + 2 2 = 2 22 = 2 32 + 2 23 + 2 33 + 2 v (2) + 2 v (3) = 1; (4.6) 2 v (3) (3) (2) 2 y (3) 2 3; + = 1: This requires 2 (2) =1 2 22 2 23 2 2 > 0 and 2 (3) =1 2 32 2 33 2 2 (2) 2 3 > 0: (4.7) In addition to (4.4), (4.7) implies two further inequality restrictions on the nine parameters of the data generating process, which are f 2; 3; ; 22 ; 23 ; 32 ; 33 ; 2; 3 g: (4.8) However, more restrictions should be respected as we will see, when we consider further consequences of a choice of particular values for these DGP parameters. 4.2. Simulation design parameter space Assigning a range of reasonable values to the nine DGP parameters is cumbersome as it is not immediately obvious what model characteristics they imply. Therefore, we now …rst de…ne econometrically meaningful design parameters. These are functions of the DGP parameters, and we will invert these functions in order to …nd solutions for the parameters of the DGP in terms of the chosen design parameter values. Since the DGP is characterized by nine parameters, we should de…ne nine variation free design parameters as well. However, their relationships will be such, that this will not automatically imply the existence nor the uniqueness of solutions. Two obvious design parameters are the degree of simultaneity in y (2) and y (3) ; given by (j) (4.9) j = Cov(yi ; ui )=( y (j) u ) = j ; j = 2; 3: Hence, by choosing 2y(2) = 2y(3) = 1, the degree of simultaneity in y (j) is directly controlled by j for j = 2; 3; and it implies two more inequality restrictions, namely j < 1; j = 2; 3: 17 (4.10) Another design parameter is a measure of multicollinearity between y (2) and y (3) given by the correlation 23 = 22 32 + 23 33 2 22 + (1 2 2) 2 23 + 2 3; (4.11) 2 3 < 1: (4.12) implying yet another restriction 22 32 + 2 22 + (1 23 33 2 23 2 2) + Further characterizations relevant from an econometric perspective are the marginal strength of instrument z (2) for y (j) and the joint strength of z (2) and z (3) for y (j) ; which are established by the (partial) population coe¢ cients of determination 2 j2 2 = and Rj;z23 2 j2 2 = Rj;z2 + 2 j3 ; (4.13) j = 2; 3: In the same vain, and completing the set of nine design parameters, are two similar characterizations of the …t of the equation of interest. Because the usual R2 gives complications under simultaneity, we focus on its reduced form equation y = ( 2 +( + 2+ 3 32 ) z (2) 22 3 ) (2) +( + 3 + 3 33 ) z (3) + (1 + 2 2 + 3 3 ) u: 2 23 (3) (4.14) This yields 2 y = ( 2 22 +( 2 2 3 32 ) 2 2 + + 3 ) +( (2) + 2 2 23 + 3 33 ) 2 2 (3) + (1 + 2 2 3 + 2 3 3) , (4.15) and in line with (4.13) we then have 2 R1;z2 = ( 2 22 + 3 32 )2 = 2y and 2 R1;z23 = [( 2 22 + 3 32 )2 + ( 2 23 + 2 2 3 33 ) ]= y : (4.16) The 9-dimensional design parameter space is given now by f 2; 3; 2 2 2 2 2 2 23 ; R2;z2 ; R2;z23 ; R3;z2 ; R3;z23 ; R1;z2 ; R1;z23 g: (4.17) The …rst three of these parameters have domain ( 1; +1) and the six R2 values have to obey the restrictions 2 2 < 1; j = 1; 2; 3: (4.18) 0 Rj;z2 Rj;z23 However, without loss of generality we can further restrict the domain of the nine design parameters, due to symmetry of the DGP with respect to: (a) the two regressors y (2) and y (3) in (4.1), (b) the two instrumental variables z (2) and z (3) ; and (c) implications which follow when all random variables are drawn from distributions with a symmetric density function. Due to (a) we may just consider cases where 2 2 2 3: (4.19) So, if one of the two regressors will have a more severe simultaneity coe¢ cient, it will 2 always be y (2) : Due to (b) we will limit ourselves to cases where 222 23 : Hence, if one 18 of the instruments for y (2) is stronger than the other, it will always be z (2) : On top of (4.18) this implies 2 2 R2;z2 0:5R2;z23 : (4.20) If (c) applies, we may restrict ourselves to cases where next to particular values for ( 2 ; 3 ); we do not also have to examine ( 2 ; 3 ). This is achieved by imposing + 0: In combination with (4.19) this leads to 2 3 1> j 3j 2 (4.21) 0: Solving the DGP parameters in terms of the design parameters can now be achieved as follows. In a …rst stage we can easily solve 7 of the 9 parameters, namely 9 > j = j > > > = 1=2 2 ; dj2 = 1; +1 j = 2; 3: (4.22) j2 = dj2 (Rj;z2 ) > > > > ; 2 2 Rj;z2 )1=2 ; dj3 = 1; +1 j3 = dj3 (Rj;z23 With (4.11) these give =( 23 22 32 2 3 )=(1 23 33 2 22 2 23 2 2 ): (4.23) Thus, for a particular case of chosen design parameter values, obeying the inequalities (4.18) through (4.21), we may obtain 24 solutions from (4.22) and (4.23) for the DGP parameters. However, some of these may be inadmissible, if they do not ful…ll the requirements (4.4) and (4.7). Moreover, we will show that not all of these 24 solutions necessarily lead to unique results on the distribution of the test statistics Wo ; Do and To . Finally, the remaining two parameters 2 and 3 can be solved from the pair of nonlinear equations 9 2 2 > (1 R1;z2 ) ( 2 22 + 3 32 )2 = R1;z2 [( 2 23 + 3 33 )2 > > 2 2 2 2 2 > > + (1 + 2 2 + 3 3 ) + 3 (3) + ( 2 + 3 ) ]; (2) = (1 2 )[( R1;z23 2 22 + 2 3 32 ) +( 2 23 + 3 33 ) 2 2 [(1 + ] = R1;z23 2 2 + 3 (3) + ( 2 2 + + 3 2 3 3) 2 2 ]: (4.24) Both these equations represent particular conic sections, specializing into either ellipses, parabolas or hyperbolas, implying that there may be zero up to eight solutions. However, it is easy to see that the three sub-set test statistics are all invariant with respect to : Note that u^ = [I X(X 0 PZ X) 1 X 0 PZ ](X + u) = [I X(X 0 PZ X) 1 X 0 PZ ]u and u^r = [I X(X 0 PZr X) 1 X 0 PZr ]u are invariant with respect to ; thus so are ^ 2 and ^ 2r : And • 2 is too, because y X ^ MZ Yo ^ = u^ MZ Yo ^ is, as follows from (2.22) and (2.23). Moreover, because ^ is invariant with respect to also is the numerator of 2 2 the three test statistics.17 Therefore, R1;z2 and R1;z23 do not really establish nuisance 17 2 ) > > > > > ; (2) Wu (1974) …nds this invariance result too, but his proof suggests that it is a consequence of normality of all the disturbances, whereas it holds more generally. 19 parameters, reducing the dimensionality of the nuisance parameter space to 7. Without loss of generality we may always set 2 = 3 = 0 in the simulated DGP’s. When (c) applies, not all 16 permutations of the signs of the four reduced form coe¢ cients lead to unique results for the test statistics, because of the following. If the sign of all elements of y (2) and (or) y (3) is changed, this means that in the general formulas the matrix X is replaced by XJ; where J is a K K diagonal matrix with diagonal elements +1 or 1; for which J = J 0 = J 1 : It is easily veri…ed that such a transformation has no e¤ect on the quadratic forms in y which constitute the test statistics Wo ; Do and To ; because it does not alter the space spanned by the matrices A and C of (2.24) nor that of the projection matrices used in the three di¤erent estimators of 2 : So, when changing the sign of all reduced form coe¢ cients, and at the same time the sign of all the elements of the vectors u; (2) and (3) ; the same test statistics are found, whereas the simultaneity and multicollinearity are still the same. This reduces the 16 possible permutations to 8, which we achieve by choosing d22 = 1. From the remaining 8 permutations four di¤erent couples yield similar 23 and values. We keep the four permutations which genuinely di¤er by choosing d23 = 1, and will give explicit attention to the four distinct cases 8 (1; 1; 1; 1) > > < (1; 1; 1; 1) (d22 ; d23 ; d32 ; d33 ) = (4.25) (1; 1; 1; 1) > > : (1; 1; 1; 1); when we generate the disturbances from a symmetric distribution, which at this stage we will. For the design parameters we shall choose various interesting combinations from 9 > 2 2 f0; :2; :5g > > > > 3 2 f :5; :2; 0; :2; :5g > = 2 f :5; :2; 0; :2; :5g (4.26) 23 > > 2 > Rj;z2 2 f:01; :1; :2; :3g > > > 2 R 2 f:02; :1; :2; :4; :5; :6g; j;z23 in as far as they satisfy the restrictions (4.18) through (4.21), provided they obey also the admissibility restrictions given by (4.4), (4.7) and (4.12). 5. Simulation …ndings on rejection probabilities In each of the R replications in the simulation study, new independent realizations are drawn on u; (2) and (3) . The three test statistics Wo ; Do and To will be calculated for both y (2) (then denoted as W 2 ; D2 ; T 2 ) and for y (3) (denoted W 3 ; D3 ; T 3 ) assuming the other regressor to be endogenous. These genuine sub-set tests will be compared with tests on the endogeneity of the full set. The latter are denoted W 23 ; D23 ; T 23 (these are tests involving 2 degrees of freedom), W32 ; D32 ; T32 (when y (3) is treated as exogenous) and W23 ; D23 ; T23 (when y (2) is treated as exogenous). The behavior under both the null and the alternative hypothesis will be investigated. These full-set tests are included to 20 better appreciate the special nature of the more subtle sub-set tests under investigation here. Every replication it is checked whether or not the null hypothesis is rejected by test statistic ; where is any of the tests indicated above. From this we obtain the Monte Carlo estimate 1 PR ! I (r) > c ( ) : (5.1) p = R r=1 Here I (:) is the indicator function that takes value one when its argument is true and zero when it is not. We take the standard form of the test statistics in which c ( ) is the -level critical value of the 2 distribution (with either 1 or 2 degrees of freedom) and in which 2 estimates have no degrees of freedom correction. The Monte Carlo estimator ! p estimates the actual rejection probability of asymptotic test procedure . When H 0 is true it estimates the actual type I error probability (at nominal level ) and when H 0 is false 1 ! p estimates the type II error probability, whereas ! p is then a (naive, when there are size distortions) estimator of the power function of the test in one particular argument (de…ned by the speci…c case of values of the design and matching DGP parameters). Estimator ! p follows the binomial distribution and has standard errors given by SE(! p ) = [! p (1 ! p )=R]1=2 : For R large, a 99:75% con…dence interval for the true rejection probability is CI99:75% = [! p 3 SE(! p ), ! p + 3 SE(! p )]: (5.2) We choose R = 10000; examine all endogeneity tests at the nominal signi…cance level of 5%, and take the sample size equal to n = 40 (mostly). Note that the boundary values for determining whether the actual type I error probability of these asymptotic tests di¤ers at this particular small sample size signi…cantly (at the very small level of :25%) from the nominal value 5% are :043 and :057 respectively. 5.1. At least one exogenous regressor In this subsection we examine cases where either both regressors y (2) and y (3) are actually exogenous or just y (3) is exogenous. Hence, for particular implementations of the sub-set and full-set tests on endogeneity the null hypothesis is true, but for some it is false. In fact, it is always true for the sub-set tests on y (3) in the cases of this subsection. We present a series of tables containing estimated rejection probabilities and each separate table focusses on a particular setting regarding the strength of the instruments. Every case consists of potentially four subcases; "a" stands for (d32 ; d33 ) = (1; 1) , "b" for (d32 ; d33 ) = ( 1; 1), "c" for (d32 ; d33 ) = (1; 1) and "d" for (d32 ; d33 ) = ( 1; 1). When both instruments have similar strength for y (2) and also (but probably stronger or weaker) for y (3) the identi…cation condition requires d32 6= d33 : Then only two of the four combinations (4.25) are feasible so that every case just consists of the two subcases "b" and "c". In Table 1 we consider cases with mildly strong instruments. In the …rst …ve cases both y (2) and y (3) are exogenous whereas the degree of multicollinearity changes. So in the …rst ten rows of the table, for all …ve distinct implementations of the three di¤erent test statistics examined, the null hypothesis is true. Because y (2) and y (3) are parametrized similarly here, the two sub-set test implementations are actually equivalent. The minor di¤erences in rejection probabilities stem from random variation, both in the 21 disturbances and in the single realizations of the instruments. The same holds for the two full-set implementations with one degree of freedom. For all implementations over the …rst …ve cases (both "b" and "c") Do shows acceptable size control, whereas Wo tends to underreject, whilst To overrejects. The sub-set version of Wo gets worse under multicollinearity (irrespective of the sign of 23 ), whereas multicollinearity increases the type I error probability of the full-set Wo tests. Both Do and To seem una¤ected by multicollinearity for these cases. When y (2) is made mildly endogenous, as in cases 6-10, the null hypothesis is still true for the sub-set tests W 3 ; D3 and T 3 : Their type I error probability seems virtually una¤ected by the actual values of 2 and 23 : For the sub-set tests W 2 ; D2 and T 2 the null hypothesis is false. Due to their di¤erences in type I error probability we cannot conclude much about power yet, but that they have some and that it is virtually una¤ected by 23 is clear. The next three columns demonstrate that it is essential that a full-set test exploits genuinely exogenous regressors, because if it does not it may falsely diagnose endogeneity of an exogenous regressor (but by a reasonably low probability when the regressors are uncorrelated). However, the next tests reported, which exploit the genuine exogeneity of y (3) ; demonstrate that in this case they do a much better job in detecting the endogenous nature of y (2) than the sub-set tests, provided there is (serious) multicollinearity. Here the full-set tests have the advantage of using an extra valid instrument. The e¤ects of multicollinearity can be explained as follows. Using the notation of the more general setup and auxiliary regression (2.19), the sub-set (full-set) tests test here the signi…cance of the regressors PZ Yo (PZ Yo ) in a regression already containing PZr X (PZr X = X), where Z = (Z Ye ) and Zr = (Zr Ye ): Regarding the sub-set test it is obvious that, because the space spanned by PZr X = (PZr Ye Yo Z1 ) does not change when Ye and Yo are more or less correlated, the signi…cance test of PZ Yo is not a¤ected by 23 : However, PZ Yo is a¤ected (positively in a matrix sense) when Yo and Ye are more (positively or negatively) correlated, which explains the increasing probability of detecting endogeneity by the present full-set tests. Finally the two degrees of freedom full-set tests demonstrate power, also when the null hypothesis tested is only partly false. One would expect lower rejection probability here than for the full-set test which correctly exploits orthogonality of y (3) ; but comparison is hampered again due to the di¤erences between type I error probabilities. Note though that the …rst …ve cases show larger type I error probabilities for T 23 than for T32 ; whereas cases 6-10 show fewer correct rejections, which fully conforms to our expectations. For a higher degree of simultaneity in y (2) (cases 11-13) we …nd for the sub-set tests that W 3 still underrejects substantially but an e¤ect of multicollinearity is no longer established, which is probably because DGP’s with a similar 2 and 3 but higher 3 3 23 are not feasible. Here D does no longer outperform T : For the other tests the rejection probabilities that should increase with j 2 j do indeed, and we …nd that the probability of misguidance by the full-set tests exploiting an invalid instrument is even more troublesome now. These results already indicate that sub-set tests are indispensable in a comprehensive sequential strategy to classify regressors as either endogenous or exogenous. Because, after a two degrees of freedom full-set test may have indicated that at least one of the two regressors is endogenous, neither of the one degree of freedom full-set tests will be capable of indicating which one is endogenous if there is one endogenous and one exogenous regressor, unless these two regressors are mutually orthogonal. However, the two sub22 set tests demonstrate that they can be used to diagnose the endogeneity/exogeneity of the regressors, especially when the endogeneity is serious, irrespective of their degree of multicollinearity. We shall now examine how these capabilities are a¤ected by the strength of the instruments. The results in Table 2 stem from similar DGP’s which di¤er from the previous ones only in the increased strength of both the instruments, which forces further limitations on multicollinearity, due to (4.7). Note that the size properties have not really improved. Due to the limitations on varying multicollinearity its e¤ects can hardly be assessed from this table. The rejection probabilities of false null hypotheses are larger when the maintained hypothesis is valid, whereas the tests which impose an invalid orthogonality condition become even more confusing when the genuine instruments are stronger. Multicollinearity still has an increasing e¤ect on the rejection probability of all the full-set tests, which is very uncomfortable for the implementations which impose a false exogeneity assumption. Staiger and Stock (1997) found that full-set tests have correct asymptotic size, although being inconsistent under weak instrument asymptotics. The following three tables illustrate cases in which the instruments are weak for one of the two potentially endogenous variables or for both. In the DGP’s used to generate Table 3, the instruments are weak for y (2) but strong for y (3) . So now the two sub-set tests examine di¤erent situations (even when 2 = 3 = 0) and so do the two one degree of freedom full-set tests. Especially the sub-set Wo tests and the two degrees of freedom W 23 test are seriously undersized. When the endogeneity of the weakly instrumented regressor is tested by W32 the type I error probability is seriously a¤ected by (lack of) multicollinearity. All full-set To tests are oversized. Only the Do tests would require just a (mostly) moderate size correction. The probability that sub-set test D2 will detect the endogeneity is small, which was already predicted immediately below (2.18). D23 will again provide confusing evidence, unless the regressors are orthogonal. Full set tests D32 and D23 have power only under multicollinearity. The latter result can be understood upon specializing (2.18) for this case, where the contributions with c1 and c3 disappear because Ke = 0 and L1 = 0: Using Z 0 Z = I and 2 Yo0 Yo = I we have to …nd a solution c2 satisfying c2 = o c2 = 0: Since o 6= 0 and 0 = ( 0) and the …rst column of vanishes asymptotically there is such a solution o o 2 indeed, but not if Yo0 Yo were nondiagonal. The situation is reversed in Table 4, where the instruments are weak for y (3) and strong for the possibly endogenous y (2) . Cases 23 and 24 are mirrored in cases 29 and 30. The Wo tests are seriously undersized, except W32 (building on exogeneity of y (3) ; it is not a¤ected by its weak external instruments) and W23 (provided the multicollinearity is substantial). The full-set To tests are again oversized. All Do implementations show mild size distortions. Because of the …ndings below (2.18) it is no surprise that the subset tests on y (2) exhibit deminishing power for more severe multicollinearity. After size correction it seems likely that W 2 or especially T 2 would do better than D2 . Also the tests W32 ; D32 and T32 show power for detecting endogeneity of y (2) when the instruments are weak for exogenous regressor y (3) ; and their power increases with multicollinearity and of course with 2 : Finally we construct DGP’s in which the instruments are weak for both regressors. Given our predictions below (2.18) and because we found mixed results when the instruments are weak for one of the two regressors, not much should be expected when both 23 are a¤ected. The results in Table 5 do indeed illustrate this. The Wo tests underreject severely, To gives a mixed picture, but Do would require only a minor size correction, although it will yield very modest power. In addition to cases in which the two instruments have similar strength for y (2) and y (3) , we present a couple of cases in which this di¤ers. Note that the inequality (4.4) is now satis…ed by all four combinations in (4.25). The reason that not every case in Table 6 consists of four subcases is that not every subcase satis…es the second part of (4.7). The results for the sub-set tests di¤er greatly between the four subcases. Subcases "a" and "d" show lower rejection probabilities for Wo and To ; whereas Do seems una¤ected under the null hypothesis. This suggests that the estimate ^ r (and hence ^ 2r ) is probably less a¤ected by (d23 ; d33 ) in these subcases than ^ 2 and • 2 : The sub-set tests on y (2) and y (3) behave similar although the (joint) instrument strength is a little higher for the former. Whereas the results between the subcases are quite di¤erent for the sub-set tests and the two degrees of freedom full-set tests, the one degree of freedom full-set test seem less dependent on the choice of (d23 ; d33 ). When y (2) is endogenous D2 has substantially less power in subcases "a" and "d" even though under the null hypothesis it rejects less often in subcases "c" and "d". For the full-set test things are di¤erent. These reject far more often in subcases "a" and "d" when there is little or no multicollinearity. However, when multicollinearity is more pronounced the tests reject less often in subcases "a" and "d" than in "b" and "c". From these results we conclude that the relevant nuisance parameters for these asymptotic tests are not just simultaneity, multicollinearity and instrument strength, but also the actual signs of the reduced form coe¢ cients. 5.2. Both regressors endogenous The rejection probabilities of the sub-set tests estimated under the alternative hypothesis in the previous subsection are only of secondary interest, because the sub-set that was treated as endogenous was actually exogenous. In such cases application of the onedegree of freedom full-set test is more appropriate. Now the not tested sub-set which is treated as endogenous will actually be endogenous, so we will get crucial information on the practical usefulness of the sub-set tests, and further evidence on the possible misguidance by the here inappropriate one degree of freedom full-set tests. Similar cases in terms of instrument strength have been chosen to keep comparability with the previous subsection. The DGP’s used for Table 7 mimic those of Table 1 in terms of instrument strength. In most cases the sub-set tests behave roughly the same as when the maintained regressor was actually exogenous, although multicollinearity is now found to have a small though clear asymmetric impact on the rejection probabilities. When multicollinearity is of the same sign as the simultaneity in y (3) , test statistics Wo and To reject less often than when these signs di¤er. This is not caused by the …xed nature of the instruments, because simulations (not reported) in which the instruments are random show the same e¤ect. On the other hand, the di¤erences between subcases diminish when the instruments are random. Multicollinearity decreases the rejection probabilities, but less so when the endogeneity of the maintained regressor is more severe. The full-set tests with one degree of freedom are a¤ected more by multicollinearity than the sub-set tests. As is to be expected, the two degrees of freedom full-set tests reject more often now that both 24 regressors are endogenous. The rejection probabilities of these full-set tests, Do included, decrease dramatically if 23 and 3 are of the same sign, and they do that much more than for the sub-set tests. Note that the cases in which 3 takes on a negative value are very similar to cases in which 3 is positive and the sign of 23 is changed, or those of (d32 ; d33 ). More speci…cally, case 63b corresponds with case 59c and case 63c with case 59b. Therefore, we will exclude cases with negative values for 3 from future tables and stick to their positive counterparts. In Table 8 we examine stronger instruments. Comparing with Table 2 we …nd that the rejection probabilities seem virtually una¤ected by choosing 3 6= 0. As we found before the rejection probabilities are a¤ected in a positive manner by the increased strength of the instruments. The sub-set tests reject almost every time if the corresponding degree of simultaneity is .5. The e¤ect of having 23 and 3 both positive seems less severe. As long as this is not the case, the two one degree of freedom full-set tests reject more often than the sub-set tests. If 23 and 3 do not di¤er in sign Wo and Do reject more often when applied to a sub-set than for their one degree and two degrees of freedom full-set versions. Because Table 3 and 4 are very similar and now both regressors are endogenous we only need to consider the equivalent table of the latter. In Table 9 the instruments are weak for y (3) but strong for y (2) . Obviously the sub-set tests for y (3) lack power now, as was already concluded from Table 3. However, sub-set tests for y (2) show power also in the presence of a maintained endogenous though weakly instrumented regressor. Note that when 3 is increased all sub-set tests for y (2) reject more often. This dependence was not apparent under non-weak instruments. As we found in Table 5 the sub-set tests perform badly when the instruments are weak for both regressors. From the results on the sub-set test for y (3) we expect the same for the case in which 3 6= 0. This we found to be true in further simulations, though we do not present a table on these as it is not very informative. These simulations demonstrate that the sub-set tests are indispensable when there is more than one regressor in a model that might be endogenous. Using only fullset tests will not enable to classify the individual variables as either endogenous or exogenous. However, all tests examined here show substantial size distortions in …nite samples. Moreover, these size distortions are found to be determined in a complex way by the model characteristics. In fact the various tables illustrate that it are not just the design parameters simultaneity, multicollinearity and instrument strength which determine the size of these tests. The di¤erences between the subcases illustrate that the size also depends on the actual reduced form coe¢ cients and therefore in fact on the degree by which the multicollinearity stems from correlation between the reduced form disturbances ( ): Trying to mitigate the size problems by simple degrees of freedom adjustments or by transformations to F statistics seems therefore a dead-end. 6. Results for bootstrapped tests Because all the test statistics that are under investigation here are based on appropriate …rst order asymptotics, it should be feasible to mitigate the size problems by bootstrapping. 25 6.1. A bootstrap routine for sub-set DWH test statistics Bootstrap routines for testing the orthogonality of all possibly endogenous regressors have previously been discussed by Wong (1996). Implementation of these bootstrap routines is relatively easy due to the fact that no regressors are assumed to be endogenous under the null hypothesis. This in contrast to the test of sub-sets where some regressors are endogenous also under the null hypothesis. Their presence complicates matters as bootstrap realizations have to be generated on both the dependent variable and the maintained set of endogenous regressors. We discuss two routines; …rst a parametric and next a semiparametric bootstrap. For the former routine we have to assume a distribution for the disturbances, which we choose to be the normal. Consider the n (1 + Ke ) matrix U = (u Ve ): Its elements can be estimated by: u^r = y X ^ r and V^er = Ye Zr ^ er , where ^ er = (Zr0 Zr ) 1 Zr0 Ye . Under the null hypothesis ^ r and ^ er are consistent estimators and it follows that U^r = (^ ur V^er ) is 1 0 consistent for U , and hence ^ = n U^r U^r is a consistent estimator of the variance of its rows. The following illustrates the steps that are required for the bootstrap procedure. 1. Draw pseudo disturbances of sample size n from the N (0; ^ ) distribution and (b) collect them in U (b) = (u(b) Ve ) . Obtain bootstrap realizations on the endogenous (b) (b) explanatory variables and the dependent variable through: Ye = Zr ^ er + Ve (b) and y (b) = X (b) ^ r + u(b) , where X (b) = (Ye Yo Z1 ). Calculate the test statistic of choice and store its value ^ (b) . 2. Repeat step (1) B times resulting in the B 1 vector ^ B = ( ^ (1) ::: ^ (B) )0 of which the elements should be sorted in increasing order. 3. The null hypothesis should be rejected if for the empirical value ^ ; calculated on the basis of y; X and Z; one …nds ^ > ^ bc , the (1 )(B + 1)-th value of the sorted vector. Applying the semiparametric bootstrap is very similar as it only di¤ers from the parametric one in step (1). Instead of assuming a distribution for the disturbances we resample by drawing rows with replacement from U^r . 6.2. Simulation results for bootstrapped test statistics Wong (1996) concludes that bootstrapping the full-set test statistics yields an improvement over using …rst order asymptotics, especially in the case where the (in his case external) instrument is weak. In this subsection we will discuss simulation results for the bootstrapped counterparts of the various test statistics. Again all results are obtained with R = 10000 and n = 40, additionally we choose the number of bootstrap replications to be B = 199. To mimic as closely as possible the way the bootstrap would be employed in practice, for each case and each test statistic we calculated the bootstrap critical value ^ bc again in each separate replication. Table 10 is the bootstrapped equivalent of Table 1. Whereas we found that the crude asymptotic version of Wo underrejects while To overrejects, bootstrapping these test statistics results in a substantial improvement18 of their size properties. In fact, 18 Although the current implementation of the bootstrap already performs quite well, even better results may be obtained by rescaling the reduced form residuals by a loss of degrees of freedom correction. 26 in this respect all three tests perform now equally well with mildly strong instruments, because the estimated actual signi…cance level lies always inside the 99.75% con…dence interval for the nominal level. Not only the sub-set tests pro…t from being bootstrapped, the one degree and two degrees of freedom full-set tests do as well. In terms of power we …nd that the bootstrapped versions of Wo ; To and Do perform almost equally well. We do …nd minor di¤erences in rejection frequencies under the alternative, but often these seem still to be the results of minor di¤erences in size. Nevertheless, on a few occasions test Do seems to fall behind. Now we establish more convincingly that exploiting correctly the exogeneity of y (2) in a full-set test provides more power, especially when multicollinearity is present, than not exploiting it in a sub-set test. Of course, the unfavorable substantial rejection probability of the exogeneity of the truly exogenous y (3) ; caused by wrongly treating y (2) as exogenous in a full-set test, cannot be healed by bootstrapping. Similar conclusions can be drawn from Table 11 which contains results for stronger instruments. On the other hand, we …nd in Table 12 that bootstrapping does not achieve satisfactory size control for most of the sub-set tests, when the instruments are weak for one regressor. Only D2 shows reasonable type I error probabilities, but when testing the endogeneity of y (2) ; the regressor for which the instruments are weak, there is hardly any power. The full-set tests do not show substantial size distortions and the one degree of freedom full-set test on y (2) and the two degrees of freedom test demonstrate power provided the regressors show multicollinearity. The results in Table 13 indicate that the sub-set test is of more use when weakness of instruments does not concern the variable under test. We can conclude that Wo and To have more power than Do ; since they reject less often under the null hypothesis but more often under the alternative. Because we were unable yet to properly size correct the sub-set test on the strongly instrumented regressor in Tables 12 and 13, we know that we will be unable to do so too when all regressors are weakly instrumented. This is supported by the results summarized in Table 14. Again the results are slightly better for Wo and To but there is almost no power. For DGP’s in which both regressors are endogenous we again construct three tables. From subsection 5.2 we learned that under the alternative hypothesis the tests behave similar to cases in which only y (2) is endogenous. This is found here too as can be seen from Table 15. We …nd further evidence that the sub-set version of Do performs less than Wo and To . New in comparison with Table 7 is that the two degrees of freedom full-set tests generally exhibit more power than the one degree of freedom full-set tests when the instruments are mildly strong. However, this was already found for cases with stronger instruments without bootstrapping. Increasing the instrument strength raises the rejection probabilities as before as can be seen from Table 16. That our current implementation of the bootstrap does not o¤er satisfactory size control for most sub-set tests when y (3) is weakly instrumented was already demonstrated in Table 12 and we conclude the same for the case when both regressors are endogenous as is obvious from the results in Table 17. 7. Empirical case study A classic application involving more than one possibly endogenous regressor is Griliches (1976), which studies the e¤ect of education on wage. It is often used to demonstrate 27 instrumental variable techniques. Both education and IQ are presumably endogenous due to omitted regressors. However, testing this assumption is often overlooked. Here we shall examine the exogeneity status of both regressors jointly and individually by means of the full-set tests and the sub-set tests. The same data are used as in Hayashi (2000, p.236). We have the wage equation and reduced form equations log Wi = 1 Si + 2 IQi + Z1i 1 + ui Yi = Z1i 1 + Z2i 2 + Vi ; (7.1) (7.2) where W is the hourly wage rate, S is schooling in years and IQ is a test score. All regressors that are assumed to be predetermined or exogenous are included in Z1 ; these are an intercept (CON S), years of experience (EXP R), tenure in years (T EN ), a dummy for southern states (RN S) and a dummy for metropolitan areas (SM SA). Additionally Z2 includes instruments age, age squared, mother education, KWW test score and a marital status dummy. In accordance with our previous notation both potentially endogenous regressors are included in Y . Table 18 presents the results of four regressions. OLS treats both schooling and IQ as exogenous, whereas they are assumed to be endogenous in the IV regression. In IV1 regressor Si is treated as predetermined and IQ as endogenous whereas in IV2 regressor IQ is treated as predetermined and schooling as endogenous. Next, in Table 19, we test various hypotheses regarding the exogeneity of one or both potentially endogenous regressors. Joint exogeneity of schooling and IQ is rejected. Hence, at least one of these regressors is endogenous and we should use the sub-set tests to …nd out whether it is just one or both. However, …rst we examine the e¤ect of using the full-set test on the individual regressors. In both cases the null hypothesis is rejected. From the Monte Carlo simulation results we learned that the full-set tests are inappropriate for correctly classifying individual regressors in the presence of other endogenous regressors. Therefore, we better employ the sub-set tests. Again we reject the null hypothesis that schooling is exogenous, but the null hypothesis that IQ is exogenous is not rejected at usual signi…cance levels. Bootstrapping these two test statistics does not lead to di¤erent conclusions. Based on these results one could greet regression IV2 instead of IV , resulting in reduced standard errors and a less controversial result on the e¤ect of IQ, as can be seen from Table 18. 8. Conclusions In this study various tests on the orthogonality of arbitrary subsets of explanatory variables are motivated and their performance is compared in a series of Monte Carlo experiments. We …nd that genuine sub-set tests play an indispensable part in a comprehensive sequential strategy to classify regressors as either endogenous or exogenous. Full-set tests have a high probability to classify an exogenous regressor wrongly as endogenous if it is merely correlated with an endogenous regressor. Regarding type I error performance we …nd that sub-set tests bene…t from estimating variances under the null hypothesis (Do ), as in Lagrange multiplier tests. Estimating the variances under the alternative (Wo ), as in Wald-type tests, leads to underrejection when the instruments are not very strong. However, bootstrapping results in good size control for all test statistics as long as the instruments are not weak for one of 28 the endogenous regressors. When the various tests are compared in terms of power the bootstrapped Wald-type tests behave often more favorable. This falsi…es earlier theoretical presumptions on the better power of the To type of test. The outcome is such that we do not expect that a better performance could have been obtained by the computationally more involved implementations that result from strictly employing the Hausman or the Hansen-Sargan principles. Even when the instruments are weak for the maintained endogenous regressor but strong for the regressor under inspection we …nd that the auxiliary regression tests exhibit power, but there is insu¢ cient size control, also when bootstrapped. This is in contrast to situations in which the instruments are not weak. Then, when bootstrapped, the sub-set and full-set tests can jointly be used fruitfully to classify individual explanatory variables and groups of them as either exogenous or endogenous. It must be noted though that the conclusions obtained from the experiments in this study are limited, as they only deal with static linear models with Gaussian disturbances, which are just identi…ed by genuinely exogenous external instruments. Apart from relaxing some of these limitations in future work, we plan to look further into e¤ects due to weakness of instruments. Furthermore, tests on the orthogonality of sub-sets of external instruments and joint tests on the orthogonality of included and excluded instruments deserve further examination. References Ahn, S.C., 1997. Orthogonality tests in linear models. Oxford Bulletin of Economics and Statistics 59, 183-186. Baum, C., Scha¤er, M., Stillman, S., 2003. Instrumental variables and GMM: estimation and testing, Stata Journal 3, 1-31. Chmelarova, V., Hill, R.C., 2010. The Hausman pretest estimator. Economics Letters 108, 96-99. Davidson, R., MacKinnon, J.G., 1989. Testing for consistency using arti…cial regressions. Econometric Theory 5, 363-384. Davidson, R., MacKinnon, J.G., 1990. Speci…cation tests based on arti…cial regressions. Journal of the American Statistical Association 85, 220-227. Davidson, R., MacKinnon, J.G., 1993. Estimation and Inference in Econometrics. New York, Oxford University Press. Doko Tchatoka, F., 2014. On bootstrap validity for speci…cation tests with weak instruments. Forthcoming in The Econometrics Journal. Durbin, J., 1954. Errors in variables. Review of the International Statistical Institute 22, 23-32. Godfrey, L.G., Hutton, J.P., 1994. Discriminating between errors-in-variables/simulaneity and misspeci…cation in linear regression models. Economics Letters 44, 359-364. Griliches, Zvi., 1976. Wages of very young men. Journal of Political Economy 84, S69-S85. Guggenberger, P., 2010. The impact of a Hausman pretest on the asymptotic size of a hypothesis test. Econometric Theory 26, 369-382. Hahn, J., Ham, J., Moon, H.R., 2011. The Hausman test and weak instruments. Journal of Econometrics 160, 289-299. 29 Hausman, J.A., 1978. Speci…cation tests in econometrics. Econometrica 46, 12511271. Hausman, J.A., Taylor, W.E., 1981. A generalized speci…cation test. Economics Letters 8, 239-245. Hayashi, Fumio, 2000. Econometrics. Princeton, Princeton University Press. Holly, A., 1982. A remark on Hausman’s speci…cation test. Econometrica 50, 749759. Hwang, Hae-shin, 1980. Test of independence between a subset of stochastic regressors and disturbances. International Economic Review 21, 749-760. Hwang, Hae-shin, 1985. The equivalence of Hausman and Lagrange multiplier tests of independence between disturbance and a subset of stochastic regressors. Economics Letters 17, 83-86. Jeong, J., Yoon, B.H., 2010. The e¤ect of pseudo-exogenous instrumental variables on Hausman test. Communications in Statistics: Simulation and Computation 39, 315321. Kiviet, J.F., 1985. Model selection test procedures in a single linear equation of a dynamic simultaneous system and their defects in small samples. Journal of Econometrics 28, 327-362. Kiviet, J.F., 2012. Monte Carlo Simulation for Econometricians. Forthcoming in: Foundations and Trends in Econometrics, Now Publishers, New York. Meepagala, G., 1992. On the …nite sample performance of exogeneity tests of Revankar, Revankar and Hartley and Wu-Hausman. Econometric Reviews 11, 337 - 353. Nakamura, A., Nakamura, M., 1981. On the relationships among several speci…cation error tests presented by Durbin, Wu, and Hausman. Econometrica 49, 1583-88. Nakamura, A., Nakamura, M., 1985. On the performance of tests by Wu and by Hausman for detecting the Ordinary Least Squares bias problem. Journal of Econometrics 29, 213-227. Newey, W., 1985. Generalized method of moments speci…cation testing. Journal of Econometrics 29, 229-256. Pesaran, M.H., Smith, R.J., 1990. A uni…ed approach to estimation and orthogonality tests in linear single-equation econometric models. Journal of Econometrics 44, 41-66. Revankar, N.S., 1978. Asymptotic relative e¢ ciency analysis of certain tests of independence in structural systems. International Economic Review 19, 165-179. Revankar, N.S., Hartley, 1973. An independence test and conditional unbiased predictions in the context of simultaneous equation systems. International Economic Review 14, 625-631. Ruud, P.A., 1984. Tests of speci…cation in econometrics. Econometric Reviews 3, 211–242 and 269–276. Ruud, P.A., 2000. An Introduction to Classical Econometric Theory. Oxford University Press. Sargan, J.D., 1958. The estimation of economic relationships using instrumental variables. Econometrica 26, 393-415. Smith, R.J., 1983. On the classical nature of the Wu-Hausman stistics for the independence of stochastic regressors and disturbance. Economics Letters 11, 357-364. Smith, R.J., 1984. A note on Likelihood Ratio tests for the independence between a subset of stochastic regressors and disturbances. International Economic Review 25, 30 263-269. Smith, R.J., 1985. Wald tests for the independence of stochastic variables and disturbance of a single linear stochastic simultaneous equation. Economics Letters 17, 87-90. Spencer, D.E., Berk, K.N., 1981. A limited information speci…cation test. Econometrica 49, 1079-1085 (and Erratum in Econometrica 50, 1087). Staiger, D., Stock, J.H., 1997. Instrumental variables regression with weak instruments. Econometrica 65, 557-586. Thurman, W.N., 1986. Endogeneity testing in a supply and demand framework. The Review of Economics and Statistics 68, 638-646. Wong, Kafu, 1996. Bootstrapping Hausman’s exogeneity test. Economics Letters 53, 139-143. Wu, D.-M., 1973. Alternative tests of independence between stochastic regressors and disturbances. Econometrica 41, 733-750. Wu, D.-M., 1974. Alternative tests of independence between stochastic regressors and disturbances: …nite sample results. Econometrica 42, 529-546. Wu, D.-M., 1983. A remark on a generalized speci…cation test. Economics Letters 11, 365-370. 31 Table 1: One endogenous regressor and mildly strong instruments: 2 2 2 2 R2;z2 = :2, R2;z23 = :4, R3;z2 = :2, R3;z23 = :4 0 0 -.2 -.2 .2 .2 -.5 -.5 .5 .5 W3 .033 .036 .032 .034 .034 .035 .024 .024 .026 .025 D3 .054 .056 .055 .057 .056 .057 .058 .056 .060 .057 T3 .064 .069 .064 .067 .065 .069 .069 .067 .070 .066 W2 .030 .030 .029 .031 .028 .029 .024 .023 .023 .023 D2 .050 .053 .055 .053 .051 .053 .057 .056 .056 .056 T2 .061 .063 .064 .062 .061 .063 .068 .066 .067 .067 W23 .040 .042 .044 .044 .045 .047 .057 .059 .054 .057 D23 .050 .050 .050 .050 .049 .052 .052 .052 .048 .052 T23 .070 .073 .069 .074 .069 .073 .073 .076 .071 .075 W32 .040 .037 .043 .040 .041 .041 .058 .054 .053 .055 D32 .046 .047 .047 .048 .046 .046 .052 .050 .048 .049 T32 .068 .069 .067 .068 .068 .068 .073 .072 .071 .070 W 23 .021 .023 .023 .024 .023 .023 .027 .029 .027 .028 D23 .044 .046 .044 .047 .044 .044 .046 .047 .045 .047 T 23 .088 .086 .090 .086 .087 .086 .090 .090 .083 .086 0 0 0 0 0 0 0 0 0 0 0 0 -.2 -.2 .2 .2 -.5 -.5 .5 .5 .033 .036 .032 .032 .034 .034 .025 .026 .027 .025 .056 .058 .057 .059 .058 .059 .059 .058 .060 .058 .064 .067 .064 .067 .064 .067 .067 .066 .071 .066 .122 .125 .120 .122 .117 .122 .103 .092 .093 .102 .177 .180 .174 .174 .168 .173 .155 .143 .144 .154 .199 .203 .198 .198 .197 .198 .192 .183 .182 .192 .039 .043 .068 .075 .076 .072 .701 .705 .709 .707 .047 .051 .076 .085 .085 .082 .682 .687 .691 .690 .074 .074 .108 .117 .116 .112 .743 .744 .754 .748 .134 .135 .172 .177 .177 .172 .759 .763 .767 .763 .161 .157 .186 .192 .195 .190 .744 .745 .751 .746 .208 .205 .239 .243 .246 .243 .798 .800 .804 .800 .067 .067 .083 .088 .090 .090 .609 .609 .614 .612 .128 .128 .147 .148 .150 .150 .644 .645 .651 .646 .200 .202 .228 .233 .231 .230 .747 .748 .755 .753 0 0 0 0 0 0 0 0 -.2 -.2 .2 .2 .028 .031 .028 .030 .031 .030 .063 .065 .064 .065 .065 .066 .059 .064 .060 .064 .062 .063 .816 .810 .768 .763 .762 .771 .865 .865 .818 .809 .808 .816 .885 .888 .862 .850 .852 .859 .041 .042 .322 .333 .335 .321 .051 .052 .344 .355 .358 .343 .075 .075 .410 .421 .425 .406 .825 .822 .929 .930 .934 .932 .848 .846 .933 .934 .938 .936 .886 .884 .952 .954 .958 .956 .634 .638 .814 .820 .818 .815 .782 .783 .895 .898 .898 .900 .861 .858 .939 .941 .944 .943 Case 1b 1c 2b 2c 3b 3c 4b 4c 5b 5c 2 3 23 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 6b 6c 7b 7c 8b 8c 9b 9c 10b 10c .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 11b 11c 12b 12c 13b 13c .5 .5 .5 .5 .5 .5 Table 2: One endogenous regressor and stronger instruments: 2 2 2 2 R2;z2 = :3, R2;z23 = :6, R3;z2 = :3, R3;z23 = :6 Case 14b 14c 15b 15c 16b 16c 2 3 23 0 0 0 0 0 0 0 0 0 0 0 0 0 0 -.2 -.2 .2 .2 W3 .048 .052 .045 .048 .048 .046 D3 .058 .061 .058 .059 .060 .058 T3 .068 .072 .068 .070 .070 .072 W2 .042 .045 .045 .045 .042 .044 D2 .052 .056 .057 .055 .054 .056 T2 .065 .068 .070 .068 .065 .068 W23 .052 .055 .054 .056 .053 .057 D23 .051 .053 .050 .052 .049 .053 T23 .070 .076 .071 .076 .070 .076 W32 .046 .048 .052 .051 .050 .051 D32 .046 .047 .048 .046 .047 .045 T32 .068 .070 .070 .070 .069 .066 W 23 .039 .042 .042 .042 .040 .042 D23 .044 .047 .045 .047 .044 .046 T 23 .089 .086 .092 .087 .087 .088 17b 17c 18b 18c 19b 19c .2 .2 .2 .2 .2 .2 0 0 0 0 0 0 0 0 -.2 -.2 .2 .2 .049 .051 .045 .046 .047 .046 .058 .061 .058 .059 .061 .059 .068 .072 .069 .071 .070 .071 .328 .329 .306 .304 .303 .307 .358 .356 .331 .329 .326 .334 .392 .391 .372 .370 .370 .375 .047 .049 .213 .220 .225 .219 .046 .047 .202 .210 .212 .208 .066 .067 .258 .260 .270 .259 .329 .328 .482 .485 .478 .480 .325 .324 .468 .471 .464 .464 .387 .388 .535 .536 .538 .534 .224 .229 .363 .361 .361 .359 .241 .247 .372 .367 .368 .365 .348 .353 .491 .488 .488 .483 20b 20c 21b 21c 22b 22c .5 .5 .5 .5 .5 .5 0 0 0 0 0 0 0 0 -.2 -.2 .2 .2 .045 .048 .043 .041 .043 .040 .063 .066 .065 .061 .061 .061 .066 .069 .067 .065 .065 .064 1 1 .994 .994 .994 .994 1 1 .992 .993 .993 .992 1 1 .996 .995 .996 .996 .023 .025 .978 .981 .980 .980 .023 .024 .975 .978 .977 .976 .037 .039 .987 .988 .987 .987 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 .999 .999 .999 .999 1 1 1 1 1 1 1 1 1 1 1 1 1 1 32 Table 3: One endogenous regressor and weak instruments for y (2) : 2 2 2 2 R2;z2 = :01, R2;z23 = :02, R3;z2 = :3, R3;z23 = :6 0 0 0 0 0 .5 0 .5 W3 .013 .014 .001 .001 D3 .031 .032 .052 .054 T3 .021 .023 .012 .016 W2 .001 0 .001 .001 D2 .054 .054 .054 .055 T2 .051 .050 .056 .057 W23 .051 .055 .056 .060 D23 .050 .055 .050 .053 T23 .072 .076 .071 .076 W32 .001 .002 .047 .051 D32 .048 .047 .049 .052 T32 .070 .070 .071 .075 W 23 .005 .005 .008 .007 D23 .044 .046 .045 .047 T 23 .086 .089 .084 .087 .2 .2 .2 .2 0 0 0 0 0 .5 0 .5 .013 .011 .003 .004 .032 .033 .053 .056 .021 .021 .026 .026 .001 0 .001 0 .057 .058 .059 .060 .053 .054 .061 .062 .050 .052 .345 .337 .048 .052 .321 .317 .071 .075 .389 .381 .001 .002 .317 .309 .047 .049 .324 .316 .072 .070 .390 .383 .005 .006 .083 .083 .047 .047 .248 .238 .089 .092 .353 .342 .5 .5 .5 .5 0 0 0 0 0 .5 0 .5 .012 .012 .023 .023 .039 .040 .060 .062 .019 .020 .126 .122 .002 .002 .002 .002 .089 .087 .091 .091 .080 .081 .103 .109 .049 .047 .069 .050 .048 .069 1 1 1 1 1 1 Case 23a 23b 24a 24b 2 3 0 0 0 0 25a 25b 26a 26b 27a 27b 28a 28b 23 .002 .062 .090 .003 .059 .086 1 1 1 1 1 1 .005 .061 .113 .005 .063 .114 .694 1 1 .696 1 1 Table 4: One endogenous regressor and weak instruments for y (3) : 2 2 2 2 R2;z2 = :3, R2;z23 = :6, R3;z2 = :01, R3;z23 = :02 0 0 0 0 0 .5 0 .5 W3 .001 .001 .001 .001 D3 .058 .058 .057 .060 T3 .054 .055 .060 .062 W2 .012 .011 .002 .001 D2 .032 .031 .056 .058 T2 .020 .021 .012 .013 W23 .002 .001 .047 .046 D23 .045 .049 .047 .048 T23 .067 .074 .070 .069 W32 .048 .049 .053 .054 D32 .046 .046 .047 .047 T32 .069 .070 .069 .068 W 23 .004 .004 .007 .006 D23 .044 .044 .044 .045 T 23 .088 .088 .087 .088 .2 .2 .2 .2 0 0 0 0 0 .5 0 .5 .001 .001 .001 .001 .059 .059 .058 .061 .056 .057 .069 .069 .100 .101 .017 .016 .137 .141 .072 .072 .141 .145 .086 .094 .006 .005 .867 .866 .156 .156 .869 .868 .201 .197 .904 .901 .328 .329 .891 .892 .324 .325 .880 .880 .386 .384 .912 .910 .068 .070 .428 .433 .242 .244 .810 .809 .347 .349 .881 .878 .5 .5 .5 .5 0 0 0 0 0 .2 0 .2 .001 .001 .001 .001 .063 .064 .063 .064 .074 .075 .074 .081 .572 .570 .414 .420 .460 .460 .298 .311 .643 .637 .536 .543 .068 .068 .429 .430 .600 .593 .870 .868 .630 .626 .883 .884 1 1 1 1 1 1 1 1 1 1 1 1 .667 .999 .658 .999 .699 1 .700 1 1 1 1 1 Case 29b 29c 30b 30c 2 3 0 0 0 0 31b 31c 32b 32c 33b 33c 34b 34c 23 Table 5: One endogenous regressor and weak instruments: 2 2 2 2 R2;z2 = R3;z2 = :01, R2;z23 = R3;z23 = :02 Case 35b 35c 36b 36c 2 3 0 0 0 0 0 0 0 0 0 .5 0 .5 W3 0 0 0 0 D3 .033 .034 .035 .033 T3 .016 .016 .020 .017 W2 0 0 .001 0 D2 .032 .031 .037 .034 T2 .017 .016 .018 .017 W23 .002 .002 .003 .002 D23 .046 .047 .046 .048 T23 .068 .070 .069 .068 W32 .002 .002 .002 .002 D32 .047 .047 .046 .048 T32 .068 .067 .065 .070 W 23 0 0 0 0 D23 .043 .044 .044 .044 T 23 .084 .083 .086 .083 37b 37c 38b 38c .2 .2 .2 .2 0 0 0 0 0 .5 0 .5 0 0 0 0 .033 .033 .037 .038 .017 .016 .019 .018 0 0 0 0 .033 .032 .036 .036 .016 .017 .018 .017 .002 .002 .003 .002 .050 .050 .049 .052 .071 .071 .074 .076 .001 .001 .002 .002 .047 .049 .050 .054 .068 .070 .075 .079 0 0 0 0 .046 .047 .050 .049 .089 .090 .095 .092 39b 39c 40b 40c .5 .5 .5 .5 0 0 0 0 0 .5 0 .5 0 0 0 0 .037 .038 .051 .050 .018 .017 .026 .024 0 .001 0 .001 .042 .041 .054 .057 .022 .021 .032 .031 .002 .002 .005 .005 .058 .057 .086 .082 .086 .084 .118 .115 .003 .003 .007 .005 .063 .061 .095 .091 .088 .089 .125 .125 0 0 0 0 .060 .059 .093 .092 .112 .112 .157 .152 23 33 Table 6: One endogenous regressor and asymmetric instrument strength: 2 2 2 2 R2;z2 = :3, R2;z23 = :5, R3;z2 = :1, R3;z23 = :4 Case 1a 1b 1c 1d 2b 2c 2d 3a 3b 3c 4c 4d 5a 5b 2 3 23 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 -.2 -.2 -.2 .2 .2 .2 -.5 -.5 .5 .5 W3 .001 .034 .033 .002 .030 .037 .002 .002 .037 .033 .032 .004 .005 .034 D3 .059 .056 .059 .060 .059 .057 .058 .058 .056 .059 .055 .054 .054 .060 T3 .036 .066 .071 .036 .069 .070 .024 .025 .068 .068 .069 .023 .026 .071 W2 .001 .037 .036 .001 .031 .038 .002 .003 .036 .030 .033 .006 .007 .030 D2 .057 .053 .052 .059 .058 .053 .057 .056 .049 .055 .054 .050 .047 .054 T2 .028 .063 .064 .028 .066 .064 .018 .018 .061 .063 .066 .022 .022 .066 W23 .057 .041 .043 .054 .049 .043 .043 .044 .041 .049 .056 .020 .022 .054 D23 .051 .050 .052 .048 .050 .051 .047 .049 .051 .050 .052 .047 .050 .049 T23 .072 .069 .073 .069 .073 .073 .069 .073 .071 .073 .075 .070 .071 .069 W32 .057 .045 .044 .054 .050 .045 .047 .048 .044 .047 .054 .035 .032 .054 D32 .050 .048 .047 .049 .048 .047 .047 .048 .046 .045 .048 .048 .046 .048 T32 .073 .070 .068 .070 .069 .071 .069 .071 .073 .067 .071 .071 .068 .069 W 23 .008 .026 .027 .008 .027 .028 .006 .008 .027 .025 .033 .006 .007 .031 D23 .048 .045 .043 .047 .045 .045 .044 .045 .044 .045 .046 .044 .044 .043 T 23 .090 .089 .086 .089 .093 .086 .089 .088 .088 .088 .089 .088 .089 .085 6a 6b 6c 6d 7b 7c 7d 8a 8b 8c 9c 9d 10a 10b .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 -.2 -.2 -.2 .2 .2 .2 -.5 -.5 .5 .5 .002 .033 .034 .003 .030 .035 .004 .003 .036 .031 .031 .005 .008 .033 .059 .057 .059 .059 .060 .059 .058 .059 .058 .060 .056 .056 .056 .061 .044 .066 .069 .046 .068 .069 .033 .032 .067 .068 .068 .032 .031 .072 .005 .190 .192 .005 .161 .203 .010 .009 .202 .162 .166 .030 .035 .164 .073 .233 .234 .072 .206 .240 .079 .080 .239 .206 .200 .104 .107 .198 .073 .264 .265 .074 .245 .272 .056 .057 .272 .245 .250 .076 .077 .248 .703 .050 .053 .710 .186 .053 .249 .246 .055 .188 .615 .056 .055 .622 .684 .059 .062 .692 .189 .062 .265 .264 .066 .192 .599 .115 .114 .608 .742 .087 .087 .751 .246 .090 .322 .322 .093 .246 .665 .151 .147 .670 .714 .224 .224 .721 .349 .231 .285 .284 .232 .343 .724 .134 .130 .729 .693 .231 .230 .700 .343 .239 .287 .286 .239 .337 .708 .172 .167 .715 .753 .289 .288 .758 .410 .301 .348 .349 .298 .406 .763 .218 .215 .771 .304 .123 .128 .310 .207 .134 .080 .082 .134 .217 .586 .037 .038 .589 .590 .177 .179 .596 .258 .185 .223 .220 .182 .266 .604 .139 .135 .607 .701 .265 .267 .706 .370 .275 .325 .317 .275 .369 .710 .214 .211 .720 11b 11c 12b 12c 12d 13a 13b 13c 14d 15a .5 .5 .5 .5 .5 .5 .5 .5 .5 .5 0 0 0 0 0 0 0 0 0 0 0 0 -.2 -.2 -.2 .2 .2 .2 -.5 .5 .032 .032 .030 .032 .014 .013 .032 .030 .025 .022 .062 .064 .065 .064 .063 .064 .063 .063 .066 .067 .064 .066 .064 .066 .089 .089 .063 .063 .074 .074 .954 .955 .853 .960 .100 .102 .961 .850 .302 .300 .958 .959 .848 .963 .201 .197 .963 .843 .432 .427 .972 .973 .911 .973 .330 .328 .975 .912 .498 .497 .145 .147 .930 .156 .989 .991 .159 .930 .438 .443 .165 .165 .930 .176 .991 .992 .179 .930 .619 .627 .208 .213 .951 .219 .995 .995 .228 .951 .674 .681 .979 .980 1 .986 .997 .997 .986 1 .867 .871 .979 .979 1 .986 .997 .996 .986 1 .909 .910 .987 .987 1 .990 .998 .998 .992 1 .934 .934 .937 .942 .996 .960 .701 .699 .963 .996 .495 .493 .960 .963 1 .971 .995 .994 .973 1 .872 .873 .981 .983 1 .987 .998 .998 .987 1 .926 .926 34 Table 7: Two endogenous regressors and mildly strong instruments: 2 2 2 2 R2;z2 = :2, R2;z23 = :4, R3;z2 = :2, R3;z23 = :4 Case 57b 57c 58b 58c 59b 59c 60b 60c 61b 61c 2 3 23 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 -.2 -.2 -.2 -.2 -.2 -.2 -.2 -.2 -.2 -.2 0 0 -.2 -.2 .2 .2 -.5 -.5 .5 .5 W3 .135 .125 .113 .106 .139 .130 .078 .072 .118 .111 D3 .192 .179 .182 .172 .182 .172 .156 .149 .150 .142 T3 .213 .202 .197 .184 .217 .204 .172 .165 .208 .201 W2 .120 .126 .109 .109 .130 .135 .079 .071 .116 .125 D2 .181 .183 .176 .177 .171 .176 .158 .146 .145 .156 T2 .203 .203 .186 .189 .207 .208 .173 .163 .202 .212 W23 .154 .142 .076 .073 .390 .382 .057 .056 1 1 D23 .177 .163 .085 .081 .414 .405 .051 .050 1 1 T23 .225 .212 .121 .113 .483 .474 .071 .072 1 1 W32 .142 .143 .073 .076 .383 .380 .057 .055 1 1 D32 .167 .168 .083 .086 .408 .405 .052 .050 1 1 T32 .215 .211 .113 .116 .478 .475 .075 .074 1 1 W 23 .138 .132 .076 .072 .338 .335 .044 .045 .999 .999 D23 .231 .229 .154 .148 .433 .426 .108 .101 1 1 T 23 .336 .327 .238 .231 .553 .546 .177 .170 1 1 62b 62c 63b 63c 64b 64c 65b 65c 66b 66c .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 0 0 -.2 -.2 .2 .2 -.5 -.5 .5 .5 .120 .131 .124 .134 .102 .112 .111 .120 .069 .075 .174 .185 .164 .177 .171 .176 .141 .148 .145 .159 .195 .207 .196 .214 .183 .192 .194 .212 .162 .173 .126 .126 .135 .131 .104 .110 .127 .114 .070 .078 .181 .182 .178 .175 .173 .176 .155 .145 .146 .158 .199 .201 .208 .208 .187 .187 .210 .203 .162 .173 .139 .149 .376 .391 .072 .075 1 1 .054 .059 .163 .173 .400 .417 .080 .086 1 1 .048 .052 .208 .221 .471 .485 .111 .117 1 1 .070 .077 .143 .143 .384 .378 .073 .074 1 1 .053 .056 .168 .166 .409 .402 .081 .084 1 1 .046 .049 .215 .214 .477 .473 .115 .114 1 1 .070 .073 .124 .132 .330 .342 .065 .068 .999 .999 .041 .045 .218 .234 .426 .431 .145 .154 1 1 .096 .103 .329 .332 .544 .547 .227 .239 1 1 .164 .177 67b 67c 68b 68c 69b 69c 70b 70c .5 .5 .5 .5 .5 .5 .5 .5 -.2 -.2 -.2 -.2 -.2 -.2 -.2 -.2 0 0 -.2 -.2 .2 .2 -.5 -.5 .137 .127 .094 .091 .159 .147 .047 .045 .210 .196 .202 .191 .195 .184 .172 .163 .216 .205 .186 .174 .229 .223 .141 .131 .811 .808 .781 .775 .746 .756 .598 .592 .857 .860 .853 .847 .767 .777 .686 .680 .879 .884 .875 .870 .830 .842 .769 .758 .194 .181 .045 .050 .882 .878 .995 .997 .221 .209 .052 .057 .891 .888 .995 .995 .272 .255 .078 .082 .919 .918 .997 .997 .848 .853 .733 .735 .996 .997 1 1 .867 .872 .750 .753 .997 .997 1 1 .899 .902 .802 .806 .998 .999 1 1 .773 .773 .586 .590 .993 .992 .987 .988 .889 .884 .766 .761 .999 .999 1 1 .936 .934 .848 .844 .999 .999 1 1 71b 71c 72b 72c 73b 73c 74b 74c .5 .5 .5 .5 .5 .5 .5 .5 .2 .2 .2 .2 .2 .2 .2 .2 0 0 -.2 -.2 .2 .2 .5 .5 .119 .131 .142 .158 .086 .094 .042 .046 .192 .202 .178 .193 .190 .196 .159 .170 .198 .210 .215 .227 .171 .177 .130 .142 .813 .806 .756 .747 .777 .779 .588 .599 .861 .855 .778 .767 .848 .853 .679 .690 .885 .880 .845 .830 .871 .873 .759 .768 .178 .189 .876 .882 .052 .049 .997 .996 .203 .216 .884 .892 .058 .056 .996 .995 .256 .270 .914 .920 .084 .081 .998 .997 .845 .846 .997 .997 .735 .734 1 1 .866 .866 .998 .997 .753 .753 1 1 .900 .897 .999 .998 .806 .802 1 1 .771 .775 .992 .993 .585 .586 .988 .987 .883 .885 .998 .998 .761 .762 1 1 .928 .935 .999 .999 .847 .846 1 1 75b 75c 76b 76c 77b 77c .5 .5 .5 .5 .5 .5 -.5 -.5 -.5 -.5 -.5 -.5 0 0 -.2 -.2 -.5 -.5 .805 .800 .803 .799 .603 .598 .838 .831 .893 .889 .804 .795 .870 .862 .895 .889 .820 .809 .794 .802 .794 .795 .599 .595 .831 .836 .886 .888 .803 .793 .862 .873 .887 .892 .816 .807 .936 .937 .416 .408 .061 .057 .947 .947 .439 .431 .055 .052 .963 .962 .504 .496 .078 .078 .939 .940 .404 .401 .061 .061 .948 .949 .427 .425 .054 .054 .963 .966 .496 .492 .081 .079 .986 .987 .849 .849 .342 .340 1 1 .972 .970 .683 .677 1 1 .985 .984 .783 .774 78b 78c 79b 79c 80b 80c .5 .5 .5 .5 .5 .5 .5 .5 .5 .5 .5 .5 0 0 .2 .2 .5 .5 .800 .809 .795 .804 .596 .603 .835 .843 .889 .892 .794 .806 .865 .872 .889 .892 .807 .816 .802 .793 .794 .793 .593 .601 .838 .828 .886 .887 .792 .799 .872 .861 .890 .887 .808 .813 .933 .935 .408 .416 .057 .060 .944 .944 .434 .440 .051 .052 .959 .961 .503 .507 .077 .078 .939 .936 .414 .407 .059 .060 .948 .946 .435 .431 .054 .054 .963 .961 .501 .496 .077 .077 .985 .987 .842 .844 .340 .346 1 1 .971 .975 .677 .684 1 1 .986 .987 .776 .783 35 Table 8: Two endogenous regressors and stronger instruments: 2 2 2 2 R2;z2 = :3, R2;z23 = :6, R3;z2 = :3, R3;z23 = :6 Case 81b 81c 82b 82c 83b 83c 2 3 23 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 0 0 -.2 -.2 .2 .2 W3 .323 .334 .305 .319 .290 .296 D3 .350 .361 .314 .328 .333 .343 T3 .384 .393 .368 .383 .358 .367 W2 .331 .330 .318 .317 .293 .296 D2 .362 .359 .326 .324 .339 .343 T2 .394 .391 .380 .378 .363 .368 W23 .332 .337 .934 .935 .100 .105 D23 .327 .332 .928 .930 .093 .098 T23 .391 .399 .948 .949 .129 .137 W32 .338 .336 .937 .936 .106 .105 84b 84c 85b 85c .5 .5 .5 .5 .2 0 .2 0 .2 .2 .2 .2 .327 .342 .273 .283 .360 .375 .357 .366 .389 .404 .349 .360 .999 .999 .999 .999 .999 .999 .999 1 .999 1 .999 .999 .402 .408 .200 .192 .395 .403 .189 .179 .475 .483 .250 .240 1 1 1 1 1 1 .998 .998 .999 .997 .997 .998 86b 86c .5 .5 .2 .5 .5 .2 1 1 .999 .999 1 1 1 1 .635 .613 .699 .637 .619 .700 .640 .622 .702 .631 .610 .692 1 1 1 1 D32 .333 .331 .932 .931 .098 .097 T32 .400 .397 .951 .952 .133 .133 W 23 .479 .481 .951 .954 .232 .244 D23 .497 .502 .948 .950 .272 .282 T 23 .618 .619 .970 .974 .384 .390 1 1 1 1 .999 1 .999 .999 1 1 1 1 1 1 1 1 1 1 W 23 .071 .072 .158 .157 .051 .050 .735 .731 .074 .074 D23 .256 .258 .435 .438 .194 .193 1 1 .224 .226 T 23 .361 .364 .556 .553 .285 .285 1 1 .326 .327 .676 1 .669 1 .731 1 .737 1 .668 .999 .662 .999 1 1 1 1 1 1 Table 9: Two endogenous regressors and weak instruments for y (3) : 2 2 2 2 R2;z2 = :3, R2;z23 = :6, R3;z2 = :01, R3;z23 = :02 Case 87b 87c 88b 88c 89b 89c 90b 90c 91b 91c 2 3 23 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 0 0 -.2 -.2 .2 .2 -.5 -.5 .5 .5 W3 0 .001 0 .001 .001 .001 .001 .001 .001 .001 D3 .060 .063 .060 .064 .061 .064 .062 .067 .060 .063 T3 .058 .060 .060 .065 .059 .061 .074 .084 .062 .066 W2 .105 .105 .107 .105 .045 .047 .042 .041 .006 .005 D2 .147 .146 .136 .131 .099 .102 .078 .075 .071 .069 T2 .147 .146 .157 .158 .077 .081 .195 .182 .038 .040 W23 .004 .005 .089 .092 .032 .030 1 1 .272 .270 D23 .162 .161 .380 .390 .172 .169 1 1 .280 .276 T23 .206 .203 .443 .450 .218 .213 1 1 .340 .336 W32 .338 .340 .547 .546 .259 .258 1 1 .311 .307 D32 .335 .334 .537 .535 .252 .250 1 1 .293 .290 T32 .398 .398 .602 .605 .311 .309 1 1 .353 .349 92b 92c 93b 93c 94b 94c .5 .5 .5 .5 .5 .5 .2 .2 .2 .2 .2 .2 0 0 -.2 -.2 .2 .2 .001 .001 0 .001 .001 .001 .065 .069 .068 .070 .066 .068 .074 .078 .083 .092 .077 .076 .584 .576 .449 .451 .407 .404 .466 .463 .304 .295 .317 .316 .655 .643 .572 .570 .526 .524 .075 .074 .477 .483 .388 .384 .613 .612 .900 .906 .829 .828 .644 .640 .910 .916 .851 .850 1 1 1 1 1 1 1 1 1 1 .999 .999 1 1 1 1 1 1 95b 95c 96b 96c 97b 97c .5 .5 .5 .5 .5 .5 .5 .5 .5 .5 .5 .5 0 0 .2 .2 .5 .5 .001 .002 .001 .002 .001 .001 .093 .101 .087 .098 .088 .092 .110 .119 .097 .106 .102 .104 .615 .605 .416 .413 .068 .071 .477 .473 .361 .359 .156 .168 .682 .671 .541 .534 .273 .285 .102 .100 .356 .348 .998 .997 .679 .675 .788 .790 .998 .997 .706 .699 .812 .815 .999 .998 1 1 .999 .999 1 .999 36 1 1 1 1 .999 .999 .999 .999 .999 1 .999 1 .702 .695 .634 .622 .676 .670 1 1 1 1 .997 .999 .997 .999 .999 .999 .998 1 Table 10: Bootstrapped: One endogenous regressor and mildly strong instruments: 2 2 2 2 R2;z2 = :2, R2;z23 = :4, R3;z2 = :2, R3;z23 = :4 0 0 -.2 -.2 .2 .2 -.5 -.5 .5 .5 W3 .053 .055 .053 .054 .052 .056 .053 .049 .052 .050 D3 .048 .052 .051 .051 .050 .052 .051 .049 .050 .049 T3 .051 .055 .052 .053 .052 .054 .052 .048 .052 .050 W2 .050 .051 .052 .051 .049 .050 .049 .047 .048 .049 D2 .046 .047 .049 .047 .046 .048 .049 .048 .048 .047 T2 .048 .050 .050 .049 .048 .049 .050 .047 .048 .048 W23 .051 .051 .052 .052 .052 .054 .053 .053 .050 .051 D23 .051 .051 .052 .052 .052 .054 .053 .053 .050 .051 T23 .051 .051 .052 .052 .052 .054 .053 .053 .050 .051 W32 .047 .049 .049 .046 .048 .046 .054 .051 .049 .047 D32 .047 .049 .049 .046 .048 .046 .054 .051 .049 .047 T32 .047 .048 .049 .046 .048 .046 .054 .051 .049 .047 W 23 .049 .049 .050 .050 .048 .049 .051 .054 .047 .051 D23 .050 .050 .050 .051 .048 .049 .049 .052 .050 .050 T 23 .050 .050 .050 .051 .048 .049 .049 .052 .050 .050 0 0 0 0 0 0 0 0 0 0 0 0 -.2 -.2 .2 .2 -.5 -.5 .5 .5 .054 .056 .052 .056 .053 .053 .051 .051 .051 .049 .049 .052 .051 .052 .051 .052 .051 .048 .051 .049 .051 .054 .051 .053 .052 .053 .052 .049 .052 .049 .176 .177 .173 .172 .169 .171 .159 .147 .150 .162 .164 .168 .162 .159 .155 .159 .138 .126 .128 .138 .172 .174 .169 .170 .166 .169 .159 .148 .150 .161 .051 .051 .079 .088 .088 .083 .680 .684 .692 .687 .051 .051 .079 .088 .088 .083 .680 .684 .692 .687 .051 .051 .079 .088 .088 .083 .680 .684 .692 .687 .162 .159 .191 .192 .195 .190 .740 .742 .750 .743 .162 .159 .191 .192 .195 .190 .740 .742 .750 .743 .162 .159 .191 .192 .195 .190 .740 .743 .750 .743 .125 .130 .150 .156 .158 .155 .697 .699 .710 .705 .132 .138 .153 .156 .160 .156 .652 .650 .657 .654 .132 .138 .153 .156 .160 .156 .652 .650 .657 .654 0 0 0 0 0 0 0 0 -.2 -.2 .2 .2 .050 .055 .050 .052 .051 .052 .051 .055 .053 .053 .054 .053 .049 .053 .048 .051 .050 .050 .864 .861 .833 .823 .827 .833 .847 .845 .788 .781 .777 .785 .861 .858 .833 .824 .825 .832 .053 .054 .342 .354 .356 .344 .053 .054 .342 .354 .356 .344 .053 .054 .342 .354 .356 .344 .844 .842 .932 .932 .938 .935 .844 .842 .932 .932 .938 .935 .844 .842 .932 .932 .938 .935 .775 .774 .896 .899 .902 .901 .788 .786 .896 .900 .901 .901 .788 .786 .896 .900 .901 .901 Case 1b 1c 2b 2c 3b 3c 4b 4c 5b 5c 2 3 23 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 6b 6c 7b 7c 8b 8c 9b 9c 10b 10c .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 11b 11c 12b 12c 13b 13c .5 .5 .5 .5 .5 .5 Table 11: Bootstrapped: One endogenous regressor and stronger instruments: 2 2 2 2 R2;z2 = :3, R2;z23 = :6, R3;z2 = :3, R3;z23 = :6 Case 14b 14c 15b 15c 16b 16c 2 3 23 0 0 0 0 0 0 0 0 0 0 0 0 0 0 -.2 -.2 .2 .2 W3 .053 .057 .051 .052 .053 .053 D3 .051 .054 .051 .050 .053 .051 T3 .052 .056 .051 .052 .053 .053 W2 0.047 0.050 0.051 0.050 0.049 0.049 D2 .045 .049 .050 .049 .049 .049 T2 .046 .049 .050 .049 .048 .049 W23 .052 .053 .052 .054 .050 .055 D23 .052 .053 .052 .054 .050 .055 T23 .052 .053 .052 .054 .050 .055 W32 .048 .048 .051 .048 .048 .046 D32 .048 .048 .051 .048 .048 .046 T32 .048 .048 .051 .048 .048 .046 W 23 .049 .050 .050 .051 .048 .049 D23 .049 .051 .050 .051 .049 .049 T 23 .049 .051 .050 .051 .049 .049 17b 17c 18b 18c 19b 19c .2 .2 .2 .2 .2 .2 0 0 0 0 0 0 0 0 .2 .2 -.2 -.2 .052 .057 .052 .051 .051 .053 .052 .056 .052 .050 .051 .052 .051 .056 .051 .050 .050 .053 0.332 0.333 0.313 0.323 0.320 0.315 .326 .328 .299 .308 .306 .298 .329 .331 .312 .320 .319 .312 .047 .047 .214 .208 .205 .208 .047 .047 .214 .208 .205 .208 .047 .047 .214 .208 .205 .208 .324 .323 .465 .464 .465 .468 .324 .323 .465 .464 .465 .468 .324 .323 .466 .464 .465 .468 .249 .255 .389 .386 .389 .387 .254 .258 .378 .376 .382 .379 .254 .258 .378 .376 .382 .379 20b 20c 21b 21c 22b 22c .5 .5 .5 .5 .5 .5 0 0 0 0 0 0 0 0 .2 .2 -.2 -.2 .050 .055 .049 .049 .050 .049 .052 .055 .049 .049 .053 .051 .050 .054 .050 .048 .050 .049 1 1 .995 .995 .994 .994 1 1 .989 .989 .989 .989 1 1 .995 .994 .994 .994 .025 .025 .976 .976 .974 .975 .025 .025 .976 .976 .974 .975 .025 .025 .976 .976 .974 .975 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 37 .999 .999 .999 .999 .999 .999 1 1 1 1 1 1 1 1 1 1 1 1 Table 12: Bootstrapped: One endogenous regressor and weak instruments for y (2) : 2 2 2 2 R2;z2 = :01, R2;z23 = :02, R3;z2 = :3, R3;z23 = :6 0 0 0 0 0 .5 0 .5 W3 .034 .037 .014 .016 D3 .038 .038 .046 .049 T3 .033 .036 .024 .026 W2 .029 .027 .029 .028 D2 .046 .047 .047 .045 T2 .042 .042 .045 .042 W23 .051 .053 .051 .055 D23 .051 .054 .051 .055 T23 .051 .054 .051 .055 W32 .050 .048 .050 .054 D32 .050 .048 .050 .054 T32 .050 .048 .050 .054 W 23 .050 .053 .050 .054 D23 .049 .050 .049 .051 T 23 .049 .050 .049 .051 .2 .2 .2 .2 0 0 0 0 0 .5 0 .5 .033 .034 .016 .018 .036 .038 .046 .049 .032 .035 .032 .033 .029 .033 .030 .033 .052 .051 .052 .052 .049 .047 .048 .050 .049 .054 .323 .316 .049 .054 .322 .316 .049 .054 .323 .316 .049 .052 .325 .315 .049 .052 .325 .315 .049 .052 .325 .315 .049 .051 .261 .259 .052 .052 .257 .248 .052 .052 .257 .248 .5 .5 .5 .5 0 0 0 0 0 .5 0 .5 .034 .034 .031 .033 .044 .044 .048 .050 .034 .034 .076 .072 .047 .046 .046 .048 .079 .079 .077 .077 .069 .071 .077 .083 .049 .049 .049 .050 .050 .050 1 1 1 1 1 1 Case 23b 23c 24b 24c 2 3 0 0 0 0 25b 25c 26b 26c 27b 27c 28b 28c 23 .064 .064 .064 .060 .060 .060 1 1 1 1 1 1 .046 .068 .068 .050 .069 .069 .848 1 1 .847 1 1 Table 13: Bootstrapped: One endogenous regressor and weak instruments for y (3) : 2 2 2 2 R2;z2 = :3, R2;z23 = :6, R3;z2 = :01, R3;z23 = :02 0 0 0 0 0 .5 0 .5 W3 .029 .031 .033 .032 D3 .050 .050 .051 .052 T3 .045 .046 .048 .050 W2 .034 .033 .013 .013 D2 .039 .035 .051 .052 T2 .033 .032 .024 .024 W23 .047 .050 .047 .049 D23 .047 .050 .047 .049 T23 .047 .050 .047 .049 W32 .049 .049 .050 .048 D32 .049 .049 .050 .048 T32 .049 .049 .050 .048 W 23 .047 .048 .048 .047 D23 .048 .050 .049 .050 T 23 .048 .050 .049 .050 .2 .2 .2 .2 0 0 0 0 0 .5 0 .5 .031 .031 .030 .032 .051 .053 .050 .051 .047 .050 .051 .054 .177 .186 .038 .041 .141 .145 .062 .060 .178 .186 .074 .079 .157 .159 .867 .866 .157 .159 .867 .866 .157 .159 .867 .866 .320 .323 .876 .879 .320 .323 .876 .879 0.320 0.323 0.876 0.879 .252 .257 .688 .692 .253 .253 .813 .813 .253 .253 .813 .813 .5 .5 .5 .5 0 0 0 0 0 .2 0 .2 .030 .031 .030 .030 .050 .052 .052 .053 .054 .057 .052 .059 .592 .590 .388 .392 .399 .404 .230 .239 .616 .613 .451 .455 .599 .594 .871 .866 .599 .594 .871 .866 .599 .594 .871 .866 1 1 1 1 1 1 1 1 1 1 1 1 Case 29b 29c 30b 30c 2 3 0 0 0 0 31b 31c 32b 32c 33b 33c 34b 34c 23 .844 .999 .999 .842 .999 .999 .861 1 1 .861 1 1 Table 14: Bootstrapped: One endogenous regressor and weak instruments: 2 2 2 2 R2;z2 = R3;z2 = :01, R2;z23 = R3;z23 = :02 Case 35b 35c 36b 36c 2 3 0 0 0 0 0 0 0 0 0 .5 0 .5 W3 .030 .030 .031 .029 D3 .038 .040 .040 .036 T3 .029 .030 .032 .030 W2 .028 .029 .027 .026 D2 .038 .036 .043 .038 T2 .028 .029 .028 .027 W23 .047 .049 .048 .049 D23 .047 .049 .048 .049 T23 .047 .049 .048 .049 W32 .048 .048 .046 .049 D32 .048 .048 .046 .049 T32 .048 .048 .046 .049 W 23 .046 .051 .045 .047 D23 .047 .049 .049 .049 T 23 .047 .049 .049 .049 37b 37c 38b 38c .2 .2 .2 .2 0 0 0 0 0 .5 0 .5 .029 .026 .027 .030 .038 .039 .042 .043 .028 .028 .031 .030 .026 .030 .028 .026 .037 .037 .040 .040 .027 .029 .030 .028 .051 .050 .050 .055 .051 .050 .050 .055 .051 .050 .050 .055 .049 .050 .053 .056 .049 .050 .053 .056 .049 .050 .053 .056 .048 .051 .048 .052 .051 .053 .056 .054 .051 .053 .056 .054 39b 39c 40b 40c .5 .5 .5 .5 0 0 0 0 0 .5 0 .5 .029 .032 .035 .035 .042 .043 .057 .054 .031 .032 .040 .039 .033 .034 .045 .040 .048 .046 .058 .062 .036 .036 .047 .047 .059 .059 .087 .086 .059 .059 .087 .086 .059 .059 .087 .086 .065 .060 .094 .092 .065 .060 .094 .092 .065 .060 .094 .092 .053 .056 .072 .073 .065 .067 .100 .098 .065 .067 .100 .098 23 38 Table 15: Bootstrapped: Two endogenous regressors and mildly strong instruments: 2 2 2 2 R2;z2 = :2, R2;z23 = :4, R3;z2 = :2, R3;z23 = :4 Case 57b 57c 58b 58c 59b 59c 60b 60c 61b 61c 2 3 23 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 -.2 -.2 -.2 -.2 -.2 -.2 -.2 -.2 -.2 -.2 0 0. -.2 -.2 .2 .2 -.5 -.5 .5 .5 W3 .186 .175 .172 .159 .191 .179 .141 .132 .173 .163 D3 .175 .166 .168 .160 .164 .154 .141 .132 .127 .120 T3 .182 .171 .169 .157 .189 .176 .140 .131 .176 .162 W2 .179 .179 .159 .162 .183 .183 .139 .129 .165 .176 D2 .169 .168 .160 .162 .155 .159 .142 .131 .123 .134 T2 .175 .175 .157 .160 .179 .179 .139 .131 .168 .177 W23 .183 .169 .087 .083 .414 .405 .052 .053 1 1 D23 .183 .169 .087 .083 .414 .405 .052 .053 1 1 T23 .183 .169 .087 .083 .414 .405 .052 .053 1 1 W32 .167 .170 .084 .085 .408 .408 .052 .051 1 1 D32 .167 .170 .084 .085 .408 .408 .052 .051 1 1 T32 .167 .170 .084 .085 .408 .408 .052 .052 1 1 W 23 .229 .227 .138 .135 .469 .453 .084 .082 1 1 D23 .242 .239 .164 .158 .447 .438 .117 .108 1 1 T 23 .242 .239 .164 .158 .447 .438 .117 .108 1 1 62b 62c 63b 63c 64b 64c 65b 65c 66b 66c .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 0 0 -.2 -.2 .2 .2 -.5 -.5 .5 .5 .171 .182 .171 .188 .158 .166 .161 .176 .128 .141 .160 .172 .149 .161 .158 .165 .119 .128 .130 .140 .167 .179 .169 .185 .155 .164 .163 .178 .128 .141 .176 .176 .184 .184 .161 .164 .175 .167 .132 .141 .167 .168 .160 .157 .160 .163 .135 .123 .133 .143 .173 .173 .180 .181 .158 .161 .179 .170 .132 .141 .165 .174 .397 .414 .082 .086 1 1 .047 .053 .165 .174 .397 .414 .082 .086 1 1 .047 .053 .165 .174 .397 .414 .082 .086 1 1 .047 .053 .169 .165 .409 .399 .083 .083 1 1 .048 .051 .169 .165 .409 .399 .083 .083 1 1 .048 .051 .169 .165 .409 .399 .083 .083 1 1 .048 .051 .220 .232 .456 .463 .128 .138 1 1 .076 .084 .233 .244 .435 .440 .152 .164 1 1 .102 .111 .233 .244 .435 .440 .152 .164 1 1 .102 .111 67b 67c 68b 68c 69b 69c 70b 70c .5 .5 .5 .5 .5 .5 .5 .5 -.2 -.2 -.2 -.2 -.2 -.2 -.2 -.2 0 0 -.2 -.2 .2 .2 -.5 -.5 .198 .185 .158 .147 .212 .205 .104 .096 .183 .170 .179 .167 .165 .153 .150 .144 .193 .179 .156 .145 .208 .200 .107 .099 .861 .863 .847 .842 .806 .815 .714 .705 .835 .836 .833 .830 .722 .729 .654 .644 .859 .861 .845 .842 .806 .817 .722 .716 .219 .209 .054 .058 .890 .883 .994 .995 .219 .209 .054 .058 .890 .883 .994 .995 .219 .209 .054 .058 .890 .883 .994 .995 .861 .866 .750 .750 .997 .997 1 1 .861 .866 .750 .750 .997 .997 1 1 .861 .866 .750 .750 .997 .997 1 1 .883 .880 .732 .737 .998 .998 .997 .997 .891 .886 .768 .771 .999 .998 1 1 .891 .886 .768 .771 .999 .998 1 1 71b 71c 72b 72c 73b 73c 74b 74c .5 .5 .5 .5 .5 .5 .5 .5 .2 .2 .2 .2 .2 .2 .2 .2 0 0 -.2 -.2 .2 .2 .5 .5 .181 .192 .199 .211 .147 .155 .097 .102 .167 .177 .148 .163 .166 .174 .141 .149 .176 .187 .195 .206 .145 .153 .098 .106 .870 .860 .820 .809 .842 .846 .708 .718 .842 .833 .734 .722 .828 .833 .644 .651 .866 .857 .822 .808 .840 .845 .716 .727 .205 .215 .882 .889 .060 .058 .995 .994 .205 .215 .882 .889 .060 .058 .995 .994 .205 .215 .882 .889 .060 .058 .995 .994 .864 .861 .997 .997 .749 .746 1 .999 .864 .861 .997 .997 .749 .746 1 .999 .864 .861 .997 .997 .749 .746 1 .999 .874 .878 .998 .998 .734 .736 .997 .997 .883 .889 .998 .998 .769 .767 1 1 .883 .889 .998 .998 .769 .767 1 1 75b 75c 76b 76c 77b 77c .5 .5 .5 .5 .5 .5 -.5 -.5 -.5 -.5 -.5 -.5 0 0 -.2 -.2 -.5 -.5 .867 .856 .875 .864 .769 .759 .793 .785 .873 .864 .779 .769 .862 .854 .874 .863 .771 .762 .859 .864 .863 .868 .765 .757 .783 .791 .860 .862 .776 .766 .854 .861 .862 .867 .766 .757 .944 .945 .440 .430 .054 .052 .944 .945 .440 .430 .054 .052 .944 .945 .440 .430 .055 .052 .946 .949 .427 .422 .054 .055 .946 .949 .427 .422 .054 .055 .946 .949 .427 .422 .054 .055 .998 .998 .945 .940 .506 .501 1 1 .972 .969 .693 .685 1 1 .972 .969 .693 .685 78b 78c 79b 79c 80b 80c .5 .5 .5 .5 .5 .5 .5 .5 .5 .5 .5 .5 0 0 .2 .2 .5 .5 .862 .867 .865 .872 .754 .765 .789 .797 .864 .871 .767 .775 .858 .865 .865 .871 .758 .767 .866 .856 .867 .865 .756 .761 .789 .781 .862 .861 .766 .771 .861 .852 .867 .864 .760 .763 .940 .944 .431 .440 .053 .054 .940 .944 .431 .440 .053 .054 .940 .944 .431 .440 .053 .054 .945 .944 .432 .431 .055 .055 .945 .944 .432 .431 .055 .055 .945 .944 .432 .431 .055 .055 .998 .998 .938 .944 .498 .510 1 1 .970 .974 .685 .694 1 1 .970 .974 .685 .694 39 Table 16: Bootstrapped: Two endogenous regressors and stronger instruments: 2 2 2 2 R2;z2 = :3, R2;z23 = :6, R3;z2 = :3, R3;z23 = :6 Case 81b 81c 82b 82c 83b 83c 2 3 23 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 0 0 -.2 -.2 .2 .2 W3 .331 .340 .318 .330 .304 .308 D3 .322 .333 .282 .294 .308 .312 T3 .329 .338 .316 .329 .303 .308 W2 .339 .338 .333 .328 .308 .312 D2 .332 .331 .297 .292 .311 .315 T2 .337 .337 .330 .326 .306 .309 W23 .324 .330 .924 .927 .097 .102 D23 .324 .330 .924 .927 .097 .102 T23 .324 .330 .924 .927 .097 .102 W32 .336 .332 .928 .928 .099 .099 D32 .336 .332 .928 .928 .099 .099 T32 .336 .332 .928 .928 .099 .099 84b 84c 85b 85c .5 .5 .5 .5 .2 0 .2 0 .2 .2 .2 .2 .345 .357 .298 .307 .325 .333 .324 .331 .343 .354 .294 .305 .999 .999 .999 .999 .998 .999 .998 .998 .999 .999 .999 .999 .395 .403 .189 .181 .395 .403 .189 .181 .395 .403 .189 .181 1 1 1 1 1 1 .998 .998 .998 .996 .996 .996 86b 86c .5 .5 .2 .5 .5 .2 .999 .999 .999 .999 .999 .999 1 1 1 1 1 1 .609 .609 .608 .618 .618 .618 .618 .618 .618 .603 .603 .603 W 23 .506 .508 .954 .958 .260 .272 D23 .510 .507 .948 .952 .283 .293 T 23 .510 .507 .948 .952 .283 .293 1 1 1 1 1 1 .999 .999 .999 .999 .999 .999 1 1 1 1 1 1 Table 17: Bootstrapped: Two endogenous regressors and weak instruments for y (3) : 2 2 2 2 R2;z2 = :3, R2;z23 = :6, R3;z2 = :01, R3;z23 = :02 Case 87b 87c 88b 88c 89b 89c 90b 90c 91b 91c 2 3 23 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 .2 0 0 -.2 -.2 .2 .2 -.5 -.5 .5 .5 W3 .031 .031 .032 .032 .031 .031 .031 .035 .034 .035 D3 .053 .055 .052 .056 .053 .056 .051 .056 .053 .055 T3 .050 .050 .049 .053 .049 .050 .054 .059 .048 .051 W2 .187 .186 .159 .158 .089 .095 .056 .051 .028 .027 D2 .149 .148 .125 .119 .097 .099 .062 .060 .062 .063 T2 .187 .189 .171 .168 .095 .101 .127 .117 .048 .049 W23 .165 .162 .378 .389 .175 .168 1 1 .281 .279 D23 .165 .162 .378 .389 .175 .169 1 1 .281 .279 T23 .165 .162 .378 .389 .175 .169 1 1 .281 .279 W32 .334 .335 .535 .534 .254 .252 1 1 .292 .288 92b 92c 93b 93c 94b 94c .5 .5 .5 .5 .5 .5 .2 .2 .2 .2 .2 .2 0 0 -.2 -.2 .2 .2 .032 .032 .032 .033 .031 .033 .055 .056 .051 .056 .056 .057 .057 .058 .060 .067 .058 .055 .599 .584 .404 .408 .392 .389 .399 .401 .223 .215 .250 .252 .626 .613 .487 .484 .451 .450 .611 .611 .898 .906 .829 .827 .611 .611 .898 .906 .829 .827 .611 .611 .898 .906 .829 .827 1 1 1 1 1 1 1 1 1 1 1 1 .999 .999 .999 .999 .999 .999 .846 1 1 .837 1 1 .877 1 1 .878 1 1 .843 .999 .999 .839 .999 .999 95b 95c 96b 96c 97b 97c .5 .5 .5 .5 .5 .5 .5 .5 .5 .5 .5 .5 0 0 .2 .2 .5 .5 .041 .046 .042 .045 .044 .045 .076 .084 .072 .082 .075 .078 .082 .089 .073 .080 .076 .078 .613 .602 .411 .407 .097 .102 .399 .397 .296 .294 .132 .142 .648 .635 .468 .464 .198 .213 .678 .674 .788 .789 .998 .997 .678 .674 .788 .789 .998 .997 .678 .674 .788 .789 .998 .997 1 1 .998 .999 .999 .999 .856 .847 .811 .801 .837 .831 40 D32 .334 .335 .535 .534 .254 .251 1 1 .292 .288 1 1 .998 .999 .999 .999 T32 .334 .335 .535 .534 .254 .251 1 1 .293 .289 1 1 .998 .999 .999 .999 W 23 .260 .269 .415 .418 .204 .203 .876 .877 .237 .232 D23 .268 .266 .447 .446 .202 .203 1 1 .230 .236 1 1 .997 .997 .998 .998 T 23 .268 .266 .447 .446 .202 .203 1 1 .230 .236 1 1 .997 .997 .998 .998 Table 18: Regression results for Griliches data, n = 758 OLS log W S IQ EXPR RNS TEN SMSA CONS IV Coef. Std. Err. .0928 .0067 .0033 .0011 .0393 .0063 -.0745 .0288 .0342 .0077 .1367 .0279 3.8952 .1091 IV1 Coef. Std. Err. .1783 .0186 -.0099 .0052 .0461 .0076 -.1014 .0358 .0398 .0090 .1291 .0321 4.1049 .3552 IV2 Coef. Std. Err. .1289 .0162 -.0088 .0050 .0348 .0070 -.1096 .0341 .0394 .0086 .1475 .0305 4.6600 .3285 Coef. Std. Err. .1550 .0113 -.0017 .0013 .0495 .0068 -.0771 .0304 .0363 .0082 .1212 .0296 3.5641 .1244 Table 19: DWH tests for Griliches data Variables Test type Full-set Full-set Full-set Sub-set Sub-set Tested S, IQ S IQ S IQ Instruments Z1 ,Z2 Z1 , Z2 , IQ Z 1 , Z2 , S Z 1 , Z2 Z 1 , Z2 Test Statistics W D T 46.87 59.42 65.13 50.56 55.90 60.96 6.28 7.24 7.38 41.16 45.24 46.74 2.72 3.12 2.88 41 Critical values 2 :05 5.99 3.84 3.84 3.84 3.84 ^ bc W :05 6.27 4.69 3.12 4.46 3.28 ^ bc D :05 6.66 4.70 3.32 4.64 3.32 bc T^:05 6.78 4.77 3.37 4.52 3.54