Proof of the Gauss-Markov Theorem c Copyright 2012 Dan Nettleton (Iowa State University) Statistics 511 1/8 The Gauss-Markov Theorem Under the Gauss-Markov Linear Model, the OLS estimator c0 β̂ of an estimable linear function c0 β is the unique Best Linear Unbiased Estimator (BLUE) in the sense that Var(c0 β̂) is strictly less than the variance of any other linear unbiased estimator of c0 β. c Copyright 2012 Dan Nettleton (Iowa State University) Statistics 511 2/8 Unbiased Linear Estimators of c0 β If a is a fixed vector, then a0 y is a linear function of y. An estimator that is a linear function of y is said to be a linear estimator. A linear estimator a0 y is an unbiased estimator of c0 β if and only if E(a0 y) = c0 β ∀ β ∈ IRp c Copyright 2012 Dan Nettleton (Iowa State University) ⇐⇒ ⇐⇒ ⇐⇒ a0 E(y) = c0 β ∀ β ∈ IRp a0 Xβ = c0 β ∀ β ∈ IRp a0 X = c0 . Statistics 511 3/8 The OLS Estimator of c0 β is a Linear Estimator We have previously defined the Ordinary Least Squares (OLS) estimator of an estimable c0 β by c0 β̂, where β̂ is any solution to the normal equations X0 Xb = X0 y. We have previously shown that c0 β̂ is the same for any β̂ that is a solution to the normal equations. We have previously shown that (X0 X)− X0 y is a solution to the normal equations for any generalized inverse of X0 X denoted by (X0 X)− . Thus, c0 β̂ = c0 (X0 X)− X0 y = `0 y (where `0 = c0 (X0 X)− X0 ) so that c0 β̂ is a linear estimator. c Copyright 2012 Dan Nettleton (Iowa State University) Statistics 511 4/8 c0 β̂ is an Unbiased Estimator of an Estimable c0 β By definition, c0 β is estimable if and only if there exists a linear unbiased estimator of c0 β. It follows from slide 3 that c0 β is estimable if and only if c0 = a0 X for some vector a. If c0 β is estimable, then `0 X = c0 (X0 X)− X0 X = a0 X(X0 X)− X0 X = a0 PX X = a0 X = c0 . Thus, by slide 3, c0 β̂ = `0 y is an unbiased estimator of c0 β whenever c0 β is estimable. c Copyright 2012 Dan Nettleton (Iowa State University) Statistics 511 5/8 Proof of the Gauss-Markov Theorem Suppose d0 y is any linear unbiased estimator other than the OLS estimator c0 β̂ = `0 y. Then we know the following: 1 d 6= ` ⇐⇒ ||d − `||2 = (d − `)0 (d − `) > 0, and 2 d0 X = `0 X = c0 =⇒ d0 X − `0 X = 00 =⇒ (d − `)0 X = 00 . We need to show Var(d0 y) > Var(c0 β̂). c Copyright 2012 Dan Nettleton (Iowa State University) Statistics 511 6/8 Proof of the Gauss-Markov Theorem Var(d0 y) = Var(d0 y − c0 β̂ + c0 β̂) = Var(d0 y − c0 β̂) + Var(c0 β̂) + 2Cov(d0 y − c0 β̂, c0 β̂). Var(d0 y − c0 β̂) = Var(d0 y − `0 y) = Var((d0 − `0 )y) = Var((d − `)0 y) = (d − `)0 Var(y)(d − `) = (d − `)0 (σ 2 I)(d − `) = σ 2 (d − `)0 I(d − `) = σ 2 (d − `)0 (d − `) > 0 by (1). Cov(d0 y − c0 β̂, c0 β̂) = Cov(d0 y − `0 y, `0 y) = Cov((d − `)0 y, `0 y) = (d − `)0 Var(y)` = σ 2 (d − `)0 ` = σ 2 (d − `)0 X[(X0 X)− ]0 c = 0 by (2). c Copyright 2012 Dan Nettleton (Iowa State University) Statistics 511 7/8 Proof of the Gauss-Markov Theorem It follows that Var(d0 y) = Var(d0 y − c0 β̂) + Var(c0 β̂) > Var(c0 β̂). 2 c Copyright 2012 Dan Nettleton (Iowa State University) Statistics 511 8/8