Conditional Expectations and Linear Regressions Walter Sosa-Escudero Econ 507. Econometric Analysis. Spring 2009 March 31, 2009 Walter Sosa-Escudero Conditional Expectations and Linear Regressions ‘All models are wrong, but some are useful’ (George E. P. Box) Box, G. E. P. and Draper, N., 1987, Empirical Model-Building and Response Surfaces, Wiley, New York, p. 424. Walter Sosa-Escudero Conditional Expectations and Linear Regressions Motivation Our last attempt with the linear model So far we have assumed we ‘know’ the model and its structure. The OLS (or the GMM) estimator consistently estimates the unknown parameters. What is OLS estimating if the underlying model is completeley unknown (possibly non-linear, endogenous, heteroskedastic, etc.) We will argue that the OLS estimator provides a good linear aproximation of the (possibly non-linear) conditional expectation. Note: this lecture is highly inspired by Angrist and Pischke (2009). Walter Sosa-Escudero Conditional Expectations and Linear Regressions Conditional Expectations Revisited E(y|x) gives the expected value of y for given values of x. It provides a reasonable representation of how x alters y. If x is random, E(y|x) is a random function. LIE: E(y) = E[E(y|x)]. We need two more properties. Walter Sosa-Escudero Conditional Expectations and Linear Regressions Decomposition property: any random variable y can be expressed as y = E(y|x) + where is a random variable satifying i) E(|x) = 0, ii) E(h(x)) = 0, where h(.) is any function of x. Intuition: any variable can be‘decomposed in two parts: the conditional expectation and an orthogonoal ‘error’ term. We are not claiming E(y|x) is linear. Walter Sosa-Escudero Conditional Expectations and Linear Regressions Proof: i) E(|x) = E y − E(y|x) |x = E(y|x) − E(y|x) = 0 ii) E(h(x)) = E h(x)E(|x) = 0 Walter Sosa-Escudero Conditional Expectations and Linear Regressions Prediction property: Let m(x) be any function of x. Then E(y|x) = argmin E (y − m(x))2 . m(x) Intuition: the conditional expectation is the best prediction, where ‘best’ means minimum mean squared error. Walter Sosa-Escudero Conditional Expectations and Linear Regressions Proof: (y − m(x))2 i2 (y − E(y|x)) + (E(y|x) − m(x)) 2 2 = y − E(y|x) + E(y|x) − m(x) + 2 y − E(y|x) E(y|x) − m(x) = h Now: The first term is not affected by the choice of m(x). The third term y − E(y|x) E(y|x) − m(x) = h(x)(x), with h(x) ≡ E(y|x) − m(x), and is equal to zero, by the decomposition property. Hence, the whole expression is minimized if m(x) = E(y|x). Walter Sosa-Escudero Conditional Expectations and Linear Regressions The population linear regression For any random variable y and a vector of K random variables x, the population linear regression function is a linear funtion r(x) = x0 β, where β is a K × 1 vector that satisfies β = argmin E (y − x0 b)2 b r(x) solves a minimum mean squared linear prediction problem. Technically, r(X) is the orthogonal projection of the random variable y on the space spanned by the elements of the random vector x. This is like the population version of what we did before with the data. Walter Sosa-Escudero Conditional Expectations and Linear Regressions Orthogonality and population coefficients: if E(xx0 ) is invertible, the condition E[x(y − x0 β)] = 0 is necessary and sufficient to guarantee x0 β exists and is the unique solution to the best linear prediction problem. Corollary β = E(xx0 )−1 E(xy) Note: this the population version of the OLS estimator Walter Sosa-Escudero Conditional Expectations and Linear Regressions Sketchy proof: Let x0 β̃ be any linear predictor. E[(y − x0 β̃)2 ] = E[(y − x0 β) + x0 (β̃ − β))2 ] = E[(y − x0 β)2 ] + 2(β̃ − β)E[x(y − x0 β)] + E(x0 (β̃ − β))2 = E[(y − x0 β)2 ] + E(x0 (β̃ − β))2 ≥ E[(y − x0 β)2 ] Walter Sosa-Escudero Conditional Expectations and Linear Regressions Linear regression and conditional expectation 1) If the E(Y |X) is linear, then r(X) = E(Y |X). Proof: if E(y|x) is linear, then E(y|x) = x0 β ∗ for some K vector β ∗ . By the decomposition property E(x(y − E(y|x)) = E(x(y − x0 β ∗ )) = 0 . Solve and get β ∗ = β, hence E(y|x) = r(x) Walter Sosa-Escudero Conditional Expectations and Linear Regressions 2) r(x) is the best linear predictor of y given x in the MMSE sense: Proof: β solves the population minimum MSE linear problem. Walter Sosa-Escudero Conditional Expectations and Linear Regressions 3) r(x) is the best linear prediction of E(y|x) in the MMSE sense. Proof: see homework Walter Sosa-Escudero Conditional Expectations and Linear Regressions Go back to the starting point of this course y = x0 β + u with E(u|x) = 0, among other assumptions. Then trivially E(y|x) = r(x) = x0 β Note that we got r(X) = E(y|x) by imposing ‘structure’ on u (predeterminedness). In the population regression approach we first imposed structure on r(x) (we’ve forced it to solve a minimum prediction problem) and we got error orthogonality as a consequence. Walter Sosa-Escudero Conditional Expectations and Linear Regressions How to justify linear regression models? 1 If we are interested in E(y|x) and this is a linear function of x, then it coincides with the linear regression function. 2 r(x) provides the best linear representation of y given x. 3 r(x) provides the best linear representation of E(y|x) given x. This says that if m(x) is a non-linear function, the OLS method will be estimating consistently the parameters of r(x), who provides the best approximation of m(x) in the MMSE sense. Walter Sosa-Escudero Conditional Expectations and Linear Regressions Empirical Illustration Walter Sosa-Escudero Conditional Expectations and Linear Regressions