Conditional Expectations and Linear Regressions

advertisement
Conditional Expectations and Linear Regressions
Walter Sosa-Escudero
Econ 507. Econometric Analysis. Spring 2009
March 31, 2009
Walter Sosa-Escudero
Conditional Expectations and Linear Regressions
‘All models are wrong, but some are useful’
(George E. P. Box)
Box, G. E. P. and Draper, N., 1987, Empirical Model-Building and Response Surfaces, Wiley, New York, p. 424.
Walter Sosa-Escudero
Conditional Expectations and Linear Regressions
Motivation
Our last attempt with the linear model
So far we have assumed we ‘know’ the model and its
structure. The OLS (or the GMM) estimator consistently
estimates the unknown parameters.
What is OLS estimating if the underlying model is completeley
unknown (possibly non-linear, endogenous, heteroskedastic, etc.)
We will argue that the OLS estimator provides a good linear
aproximation of the (possibly non-linear) conditional expectation.
Note: this lecture is highly inspired by Angrist and Pischke (2009).
Walter Sosa-Escudero
Conditional Expectations and Linear Regressions
Conditional Expectations Revisited
E(y|x) gives the expected value of y for given values of x.
It provides a reasonable representation of how x alters y.
If x is random, E(y|x) is a random function.
LIE: E(y) = E[E(y|x)].
We need two more properties.
Walter Sosa-Escudero
Conditional Expectations and Linear Regressions
Decomposition property: any random variable y can be expressed
as
y = E(y|x) + where is a random variable satifying i) E(|x) = 0, ii)
E(h(x)) = 0, where h(.) is any function of x.
Intuition: any variable can be‘decomposed in two parts: the conditional
expectation and an orthogonoal ‘error’ term. We are not claiming E(y|x)
is linear.
Walter Sosa-Escudero
Conditional Expectations and Linear Regressions
Proof:
i)
E(|x) = E y − E(y|x) |x = E(y|x) − E(y|x) = 0
ii)
E(h(x)) = E h(x)E(|x) = 0
Walter Sosa-Escudero
Conditional Expectations and Linear Regressions
Prediction property: Let m(x) be any function of x. Then
E(y|x) = argmin E (y − m(x))2 .
m(x)
Intuition: the conditional expectation is the best prediction, where
‘best’ means minimum mean squared error.
Walter Sosa-Escudero
Conditional Expectations and Linear Regressions
Proof:
(y − m(x))2
i2
(y − E(y|x)) + (E(y|x) − m(x))
2 2
=
y − E(y|x) + E(y|x) − m(x)
+ 2 y − E(y|x) E(y|x) − m(x)
=
h
Now:
The first term is not affected by the choice of m(x).
The third term y − E(y|x) E(y|x) − m(x) = h(x)(x), with
h(x) ≡ E(y|x) − m(x), and is equal to zero, by the decomposition
property.
Hence, the whole expression is minimized if m(x) = E(y|x).
Walter Sosa-Escudero
Conditional Expectations and Linear Regressions
The population linear regression
For any random variable y and a vector of K random variables x,
the population linear regression function is a linear funtion
r(x) = x0 β, where β is a K × 1 vector that satisfies
β = argmin E (y − x0 b)2
b
r(x) solves a minimum mean squared linear prediction
problem.
Technically, r(X) is the orthogonal projection of the random
variable y on the space spanned by the elements of the
random vector x.
This is like the population version of what we did before with
the data.
Walter Sosa-Escudero
Conditional Expectations and Linear Regressions
Orthogonality and population coefficients: if E(xx0 ) is invertible,
the condition
E[x(y − x0 β)] = 0
is necessary and sufficient to guarantee x0 β exists and is the unique
solution to the best linear prediction problem.
Corollary
β = E(xx0 )−1 E(xy)
Note: this the population version of the OLS estimator
Walter Sosa-Escudero
Conditional Expectations and Linear Regressions
Sketchy proof: Let x0 β̃ be any linear predictor.
E[(y − x0 β̃)2 ]
= E[(y − x0 β) + x0 (β̃ − β))2 ]
= E[(y − x0 β)2 ] + 2(β̃ − β)E[x(y − x0 β)] + E(x0 (β̃ − β))2
= E[(y − x0 β)2 ] + E(x0 (β̃ − β))2
≥ E[(y − x0 β)2 ]
Walter Sosa-Escudero
Conditional Expectations and Linear Regressions
Linear regression and conditional expectation
1) If the E(Y |X) is linear, then r(X) = E(Y |X).
Proof: if E(y|x) is linear, then E(y|x) = x0 β ∗ for some K vector β ∗ . By
the decomposition property
E(x(y − E(y|x)) = E(x(y − x0 β ∗ )) = 0
. Solve and get β ∗ = β, hence E(y|x) = r(x)
Walter Sosa-Escudero
Conditional Expectations and Linear Regressions
2) r(x) is the best linear predictor of y given x in the MMSE sense:
Proof: β solves the population minimum MSE linear problem.
Walter Sosa-Escudero
Conditional Expectations and Linear Regressions
3) r(x) is the best linear prediction of E(y|x) in the MMSE sense.
Proof: see homework
Walter Sosa-Escudero
Conditional Expectations and Linear Regressions
Go back to the starting point of this course
y = x0 β + u
with E(u|x) = 0, among other assumptions. Then trivially
E(y|x) = r(x) = x0 β
Note that we got r(X) = E(y|x) by imposing ‘structure’ on u
(predeterminedness). In the population regression approach we
first imposed structure on r(x) (we’ve forced it to solve a
minimum prediction problem) and we got error orthogonality as a
consequence.
Walter Sosa-Escudero
Conditional Expectations and Linear Regressions
How to justify linear regression models?
1
If we are interested in E(y|x) and this is a linear function of
x, then it coincides with the linear regression function.
2
r(x) provides the best linear representation of y given x.
3
r(x) provides the best linear representation of E(y|x) given x.
This says that if m(x) is a non-linear function, the OLS method
will be estimating consistently the parameters of r(x), who
provides the best approximation of m(x) in the MMSE sense.
Walter Sosa-Escudero
Conditional Expectations and Linear Regressions
Empirical Illustration
Walter Sosa-Escudero
Conditional Expectations and Linear Regressions
Download