Proof of the Gauss Markov theorem
Problem
- Setup:
- a linear regression model \(y=Xβ+ε\) with a random sample
- the Gauss-Markov assumptions, \(E\left( ε \right|X)=0\) and \(Var\left( ε \right|X)=σ^2I\)
- \(b={\left( X'X \right)}^{-1}X'y\) is the OLS estimator of \(β\)
Let \(\tilde{b}=Ay\) be an arbitrary unbiased linear estimator of \(β\) where \(A\) is an arbitrary \(k×n\) matrix that may depend on \(X\) . ( \(A={\left( X'X \right)}^{-1}X'\) for the OLS estimator). Show that
\[Var\left( \tilde{b}|X \right)-Var\left( b \right|X)\]
is a positive semidefinite matrix.
Hints
a) Evaluate \(E\left( \tilde{b} \mid X \right)\) and show that unbiasedness of \(\tilde{b}\) implies that \(AX=I\)
b) Use the following neat idea (conditionally on \(X\) )
\[Var\left( \tilde{b} \right)=Var\left( b+\left( \tilde{b}-b \right) \right)\]
Now show that
\[Var\left( b+\left( \tilde{b}-b \right) \right)=\]
\[=E\left( \left( b-β \right)\left( b-β \right)' \right)+E\left( \left( b-β \right)\left( \tilde{b}-b \right)' \right)+E\left( \left( \tilde{b}-b \right)\left( b-β \right)' \right)+E\left( \left( \tilde{b}-b \right)\left( \tilde{b}-b \right)' \right)\]
c) Let’s focus on the second term, \(E\left( \left( b-β \right)\left( \tilde{b}-b \right)' \right)\) .
Before attacking this term, we have
\[\tilde{b}-b=Ay-{\left( X'X \right)}^{-1}X'y=\left( A-{\left( X'X \right)}^{-1}X' \right)y\]
Define \(C= A-{\left( X'X \right)}^{-1}X'\) such that
\[\tilde{b}-b=Cy\]
Show that \(AX=I\) implies that \(CX=0\) and \(Cy=Cε\)
d) Now show that
\[E\left( \left( b-β \right)\left( \tilde{b}-b \right)' \right)=0\]
as well as
\[E\left( \left( \tilde{b}-b \right)\left( b-β \right)' \right)=0\]
(this is a \(k×k\) matrix of zeroes). You are now actually done! You have showed that (conditionally on \(X\) )
\[Var\left( \tilde{b} \right)=Var\left( b+\left( \tilde{b}-b \right) \right)=Var\left( b \right)+Var\left( \left( \tilde{b}-b \right) \right)\]
or
\[Var\left( \tilde{b} \right)-Var\left( b \right)=Var\left( \left( \tilde{b}-b \right) \right)\]
The right-hand side, being a variance matrix, must be positive semi definite.
Solution
a)
\(E\left( \tilde{b} \mid X \right)=E\left( Ay \mid X \right)=E\left( AXβ+Aε \mid X \right)=AXβ+AE\left( ε \mid X \right)=AXβ\) .
\(E\left( \tilde{b} \mid X \right)\) must be equal to \(β\) and \(AX=I\) .
b) The expected value of \(b+\left( \tilde{b}-b \right)\) is \(β\) so (conditionally on \(X\) )
\[Var\left( b+\left( \tilde{b}-b \right) \right)=E\left( \left( b+\left( \tilde{b}-b \right)-β \right){\left( b+\left( \tilde{b}-b \right)-β \right)}' \right)=\]
\[=E\left( \left( \left( b-β \right)+\left( \tilde{b}-b \right) \right){\left( \left( b-β \right)+\left( \tilde{b}-b \right) \right)}' \right)=\]
\[E\left( \left( b-β \right)\left( b-β \right)' \right)+E\left( \left( b-β \right)\left( \tilde{b}-b \right)' \right)+E\left( \left( \tilde{b}-b \right)\left( b-β \right)' \right)+E\left( \left( \tilde{b}-b \right)\left( \tilde{b}-b \right)' \right)\]
c)
\[CX=\left( A-{\left( X'X \right)}^{-1}X' \right)X=AX-{\left( X'X \right)}^{-1}X'X=I-I=0\]
and
\[Cy=C\left( Xβ+ε \right)=CXβ+Cε=Cε\]
d) \(b-β={\left( X'X \right)}^{-1}X'ε\) and \(\tilde{b}-b=Cy=Cε\) and \(\left( \tilde{b}-b \right)'=ε'C'\) so (conditionally on X)
\[E\left( \left( b-β \right)\left( \tilde{b}-b \right)' \right)=E\left( {\left( X'X \right)}^{-1}X'εε'C' \right)={\left( X'X \right)}^{-1}X'E\left( εε' \right)C'\]
\[{\left( X'X \right)}^{-1}X'σ^2IC'=σ^2{\left( X'X \right)}^{-1}X'C'=σ^2{\left( X'X \right)}^{-1}{\left( CX \right)}'=σ^2{\left( X'X \right)}^{-1}⋅0=0\]
Take the transpose of both sides of \(E\left( \left( b-β \right)\left( \tilde{b}-b \right)' \right)=0\) :
\[E\left( \left( \tilde{b}-b \right)\left( b-β \right)' \right)=0\]