Hypothesis Testing
If we were interested in making inferences about \beta in a simple linear regression model i.e. y_i = \alpha+\beta x_i+\epsilon_i \mathbf{b}^T\boldsymbol{\beta} = \beta i.e. \mathbf{b}^T = (0 \quad 1) and this gives us:
\frac{\hat{\beta}-\beta}{\text{e.s.e}(\hat{\beta})} \sim t(n-p)
\newline Under the null hypothesis:
\newline H_0: \beta=0 (where H_1: \beta \neq 0)
\newline \frac{\hat{\beta}-0}{\text{e.s.e}(\hat{\beta})}=\frac{\hat{\beta}}{\text{e.s.e}(\hat{\beta})} \sim t(n-p)
and \frac{\hat{\beta}}{\text{e.s.e}(\hat{\beta})} is typically called the
test statistic. Therefore, the null hypothesis is rejected for large
absolute values of the test statistic, usually values > 2 i.e. for small
p-values in R
(where a p-value is the probability that we obtain a
test statistic value as extreme or more extreme if the null hypothesis is
true). In general, we reject H_0 for p-values <0.05 and this would
indicate a significant relationship between a response and an
explanatory variable in the model (i.e., \beta \neq 0).
Notice we can test any hypothsis with respect to \beta, for instance
\newline H_0: \beta=5 (where H_1: \beta \neq 5)
\newline \frac{\hat{\beta}-5}{\text{e.s.e}(\hat{\beta})} \sim t(n-p).