Hypothesis testing in the LRM: The t-test

Summary

Setup:

  • The LRM with random sampling and GM

\[y_i=β_1+β_2x_i+ε_i i=1,…,n\]

  • \(ε_i\) follows a normal distribution (conditionally on \(x_i\) ), \(ε_i|x_i∼N\left( 0,σ^2 \right) i=1,…,n\)

\(H_0: β_2=0\)

  • For the null hypothesis \(H_0: β_2=0\) we define a test statistic called the t -value or the t -statistic

\[t= \frac{b_2}{SE(b_2)}\]

  • If the null hypothesis is true then \(t∼t_{n-2}\)
  • We decide on a level of significance \(α\) and reject the null if

\[\left| t \right|>t_{α/2,n-2}\]

\(p\) -values

  • For the hypothesis above, we define the \(p\) -value of the hypothesis as the level of significance \(α\) such that we are indifferent between rejecting and not rejecting the null,

\[\left| t \right|=t_{p/2,n-2}\]

  • We will reject the null hypothesis if \(p<α\) .
  • We will not reject the null hypothesis if \(p>α\) .
  • “If \(p\) is low the null must go. If \(p\) is high the null will fly”.

Large \(n\)

  • If the GM assumptions are satisfied but the errors \(ε_i\) do not follow a normal distribution then, under mild conditions, the \(t\) -statistic will converge to a standard normal distribution as \(n→∞\) . This follows by a result known as the central limit theorem (CLT).
  • Therefore, if \(n\) is large, the procedures outlined in this section will be approximately correct even if the errors are not normal.