Summary points

In lecture 4, we

  • Defined a linear model in vector-matrix notation \mathbf{Y} = \mathbf{X}\boldsymbol{\beta}+\boldsymbol{\epsilon}.

  • Construted the formula for least squares estimators \boldsymbol{\hat{\beta}} = (\mathbf{X}^T\mathbf{X})^{-1}\mathbf{X}^T\mathbf{Y}.

  • Derived the residual sum of squares RSS=\mathbf{Y}^T\mathbf{Y}-\mathbf{Y}^T\mathbf{X}\boldsymbol{\hat{\beta}}.

These formulae cover all least-squares problems, for suitably defined \mathbf{X} and \boldsymbol{\beta} they are worth remembering. To solve a particular problem, you just need to plug in the correct \mathbf{X}. For any example, we simply need to identify the matrix \mathbf{X} and the vectors \mathbf{Y} and \boldsymbol{\beta} and then apply these two key results. Remember that evaluating the sum of squares at the least squares estimates for \boldsymbol{\beta} gives the residual sum of squares.