Examples

Protein in pregnancy

Continuing the example from last lecture with data

Data: \((y_i,x_i) \quad i=1,\dots,19\)

Model: \(E(Y_i) = \alpha+\beta x_i\)

We now are asked to:

  1. Construct a confidence interval for \(\beta\) to test it’s significant.

  2. Predict a future observation of protein level in a health women at \(x=27\) weeks.

To answer the questions we will need the following:

\[\mathbf{X} =\left( \begin{array}{cc} 1 & x_1 \\ . & . \\ . & . \\ . & . \\ . & . \\ 1 & x_n \\ \end{array} \right)\]

\[\begin{aligned} (\mathbf{X}^T\mathbf{X}) &= \left( \begin{array}{cc} n & \sum_{i=1}^n x_i \\ \sum_{i=1}^n x_i & \sum_{i=1}^n x_i^2 \\ \end{array} \right) =\left( \begin{array}{cc} 19 & 456 \\ 456 & 12164 \\ \end{array} \right)\\ (\mathbf{X}^T\mathbf{X})^{-1} &= \left( \begin{array}{cc} 0.524763 & -0.019672\\ -0.019672 & 0.000820 \\ \end{array} \right)\quad \quad \quad \quad \quad \quad \quad (2)\\\end{aligned}\]

and the R output

pregnancy<-read.csv("week6/Lecture12/PROTEIN.CSV",header=T)
fit1<-lm(formula = Protein ~ Gestation,data=pregnancy)
summary(fit1)
## 
## Call:
## lm(formula = Protein ~ Gestation, data = pregnancy)
## 
## Residuals:
##      Min       1Q   Median       3Q      Max 
## -0.16853 -0.08720 -0.01009  0.08578  0.20422 
## 
## Coefficients:
##             Estimate Std. Error t value Pr(>|t|)    
## (Intercept) 0.201738   0.083363   2.420    0.027 *  
## Gestation   0.022844   0.003295   6.934 2.42e-06 ***
## ---
## Signif. codes:  0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1
## 
## Residual standard error: 0.1151 on 17 degrees of freedom
## Multiple R-squared:  0.7388, Adjusted R-squared:  0.7234 
## F-statistic: 48.08 on 1 and 17 DF,  p-value: 2.416e-06

Confidence Interval

So, a 95% C.I. for \(\beta\) is

\[0.0228\pm t(17;0.975)\sqrt{\frac{0.2251}{17}\mathbf{b}^T(\mathbf{X}^T\mathbf{X})^{-1}\mathbf{b}}\]

and since \(\mathbf{b} = \left( \begin{array}{c} 0 \\ 1 \\ \end{array} \right)\) ie

\[\begin{aligned} 0.0228 &\pm& 2.11\sqrt{\frac{0.2251}{17}0.000820}\\ 0.0228 &\pm& 2.11(0.003295)\end{aligned}\]

i.e. \(0.0228 \pm 0.0070\),

i.e. (0.016, 0.030)

\(\newline\) The fact that this interval contains only positive values tells us that there is clear evidence that the average level of protein increases with gestation. The coefficient for \(\beta\) is highly likely to lie somewhere between 0.02 and 0.03.

\(\newline\) Note: A confidence interval that includes zero indicates that there insufficient evidence of a relationship between the response and the explanatory variable. In this situation we would expect the p-value for testing H\(_0:\beta=0\) to be \(>0.05\).

Prediction Interval

In order to use this model in a clinical setting we need a means of telling what values of protein level are expected for a future healthy mother who attends this clinic. For example, if a woman who is 27 weeks pregnant has a protein level of 1.06, should this be regarded as unusual? A prediction interval helps us to answer this question.

A 95% P.I. for a future observation \(y\) at \(x = 27\) is done in the following way:

Here \(\mathbf{b} = \left( \begin{array}{c} 1 \\ 27 \\ \end{array} \right)\).

Thus

\[\begin{aligned} \boldsymbol{b}^T(\boldsymbol{X}^T\boldsymbol{X})^{-1}\boldsymbol{b} &= \left(\begin{array}{cc} 1 & 27 \end{array}\right) \left(\begin{array}{cc} 0.5247 & -0.0196 \\ -0.0196 & 0.0008 \end{array}\right) \left(\begin{array}{c} 1 \\ 27 \end{array}\right)\\ &= \left(\begin{array}{cc} -0.0064& 0.0025 \end{array}\right) \left(\begin{array}{c} 1 \\ 27 \end{array}\right)\\ &= 0.060255 \end{aligned}\]

and the prediction interval is \[\begin{aligned} (0.2017 + 27 \times 0.0228) &\pm 2.11\sqrt{\frac{0.2251}{17}(1+0.060255)}\\ (0.57,&1.07)\end{aligned}\]

Since it lies within this interval (just) we have strong grounds for regarding a protein level of 1.06 as unusual. (Even if it did lie outside the interval, this only says that the result is unusual).

\(\newline\) Note that Prediction intervals will always be wider than confidence intervals.

Trees

The dataset refers to the volume (cubic feet) and diameter (inches) (at 54 inches above the ground) and height (feet) for a sample of 31 black cherry trees in the Allegheny National Forest Pennsylvania. The data were collected in order to and an estimate for the volume of a tree (and therefore for the timber yield), given its height and diameter. A starting point for estimating volume using these data is the geometric formula for a cylinder:

Our full model for the trees data is

\[E(Y) = \alpha + \beta x_1 + \gamma x_2\]

where \(Y\) denotes log(volume), \(x_1\) denotes log(diameter) and \(x_2\) denotes log(height) of 31 trees.

The fitted model produced:

\[\boldsymbol{\hat{\beta}} = \left( \begin{array}{c} -6.632 \\ 1.983 \\ 1.117 \\ \end{array} \right)\]

\[RSS = 0.1855\]

\[(\mathbf{X}^T\mathbf{X})^{-1} =\left( \begin{array}{ccc} 96.5721 & 3.1393 & -24.1651 \\ 3.1393 & 0.8495 & -1.2275 \\ -24.1651 & -1.2275 & 6.3099 \\ \end{array} \right)\]

The matrix \((\mathbf{X}^T\mathbf{X})^{-1}\) was obtained from R.

\(\newline\) Construct 95% confidence intervals for \(\beta\) and \(\gamma\) to test their significance.

In order to check whether or not there is clear evidence of a relationship between each of the explanatory variables and the response, we can construct interval estimates for \(\beta\) and \(\gamma\).

A 95% C.I. for \(\beta\) is given by

\[\mathbf{b}^T\boldsymbol{\hat{\beta}}\pm t(n-3;0.975)\sqrt{\frac{RSS}{n-3}\mathbf{b}^T(\mathbf{X}^T\mathbf{X})^{-1}\mathbf{b}}\]

where \(\mathbf{b} = \left( \begin{array}{c} 0 \\ 1 \\ 0 \\ \end{array} \right)\)

\[\hat{\beta} \pm t(28;0.975)\sqrt{\frac{0.1855}{28}0.8495}\]

\[1.983\pm 0.15\]

\[(1.83, 2.13)\]

This interval does not contain zero. There is therefore clear evidence of a relationship between log (diameter) and log (volume), i.e. log diameter is a significant predictor in addition to log height, and it is highly likely that the coefficient for log diameter lies between 1.83 and 2.13.

Note also that 2 is a plausible value for the coefficient of log (diameter). This is therefore consistent with the cylindrical model discussed in chapter 2, where (\(\mathrm{V}=\pi(\frac{\mathrm{d}}{2})^2\mathrm{h}\); log(V) = (\(\pi/4\))+2logd+logh).

A 95% C.I. for \(\gamma\) is given by

\[\mathbf{b}^T\boldsymbol{\hat{\beta}}\pm t(n-3;0.975)\sqrt{\frac{RSS}{n-3}\mathbf{b}^T(\mathbf{X}^T\mathbf{X})^{-1}\mathbf{b}}\]

where \(\mathbf{b} = \left( \begin{array}{c} 0 \\ 0 \\ 1 \\ \end{array} \right)\)

\[\hat{\gamma} \pm t(28;0.975)\sqrt{\frac{0.1855}{28}6.3099}\]

\[1.12\pm 0.42\]

\[(0.70, 1.54)\]

Again there is clear evidence of a relationship between log (height) and log (volume), since 0 does not lie in the interval estimate. Log height is a significant predictor in addition to log diameter. The results are again consistent with the cylindrical model since the value 1 lies in the interval. It is highly likely that the coefficient for log height lies between 0.70 and 1.54.

Confidence interval for the difference in two population means

Data: \((y_{ij}); \quad i=1,2; j=1,\dots,n_i\)

\(\newline\) Model: \(E(Y_{ij}) = \mu_i\), \(Y_{ij} \sim N(\mu_i, \sigma^2)\)

\(\newline\) Construct a confidence interval for \((\mu_1-\mu_2)\)

\[\mathbf{Y} = \left( \begin{array}{c} y_{11} \\ y_{12} \\ . \\ y_{1,n_1} \\ y_{21} \\ y_{22} \\ . \\ y_{2,n_2} \\ \end{array} \right) \quad \boldsymbol{\beta} = \left( \begin{array}{c} \mu_1 \\ \mu_2 \\ \end{array} \right), \quad \mathbf{X} = \left( \begin{array}{cc} 1 & 0 \\ 1 & 0 \\ . & . \\ 1 & 0 \\ 0 & 1 \\ 0 & 1 \\ . & . \\ 0 & 1 \\ \end{array} \right)\]

Interest is in \((\mu_1-\mu_2)\), i.e. \(\mathbf{b} = \left( \begin{array}{c} 1 \\ -1 \\ \end{array} \right)\)

\[(\mathbf{X}^T\mathbf{X}) = \left( \begin{array}{cc} n_1 & 0 \\ 0 & n_2 \\ \end{array} \right), \quad (\mathbf{X}^T\mathbf{X})^{-1} =\left( \begin{array}{cc} \frac{1}{n_1} & 0 \\ 0 & \frac{1}{n_2} \\ \end{array} \right)\]

\[\mathbf{X}^T\mathbf{Y} = \left( \begin{array}{c} \sum_{j=1}^{n_1}y_{1j} \\ \sum_{j=1}^{n_2}y_{2j} \\ \end{array} \right), \quad (\mathbf{X}^T\mathbf{X})^{-1}\mathbf{X}^T\mathbf{Y} =\left( \begin{array}{c} \bar{y}_1 \\ \bar{y}_2 \\ \end{array} \right) = \boldsymbol{\hat{\beta}}\]

\[\mathbf{b}^T\boldsymbol{\hat{\beta}} = (\bar{y}_1-\bar{y}_2), \quad \mathbf{b}^T(\mathbf{X}^T\mathbf{X})^{-1}\mathbf{b} = \left( \begin{array}{cc} 1 & -1 \\ \end{array} \right)\left( \begin{array}{cc} \frac{1}{n_1} & 0 \\ 0 & \frac{1}{n_2} \\ \end{array} \right)\left( \begin{array}{c} 1 \\ -1 \\ \end{array} \right) = (\frac{1}{n_1}+\frac{1}{n_2})\]

What about the RSS (residual sum-of-squares)?

\[\begin{aligned} RSS&=\mathbf{Y}^T\mathbf{Y}-\mathbf{Y}^T\mathbf{X}\boldsymbol{\hat{\beta}}\\ &=\sum_{i=1}^{n_1}\sum_{j=1}^{n_2}y_{ij}^2-(n_1\bar{y_1}^2+n_2\bar{y_2}^2)\\ &=RSS_1+RSS_2\\\end{aligned}\]

where \(RSS_i=\sum_{j=1}^{n_i}(y_{ij}-\bar{y_i})^2\)

Interval for \(\mu_1-\mu_2\):

\[(\bar{y_1}-\bar{y_2})\pm t\left(n_1+n_2-2; \frac{1+c}{2}\right)\sqrt{\frac{RSS_1+RSS_2}{n_1+n_2-2}\left(\frac{1}{n_1}+\frac{1}{n_2}\right)}\]

Autoanalyser data

Blood plasma concentrations are usually measured using a lengthy laboratory process. A simpler, cheaper method using an autoanalyser is often used. The autoanalyser is regularly tested to see if it is performing properly. On this occasion, 12 measurements have been made on samples of known concentration (3 replicates at each of 4 concentrations).

The following model has been fitted in R to estimate the autoanalyser concentration from the true concentration:

\[\mathrm{autoanalyser}_i = \beta_0 + \beta_1 \mathrm{true}_i + \epsilon_i, \quad i=1, \dots, 12\]

Construct a 95% C.I. for the population mean autoanalyser concentration when the true concentration is 6 units and interpret the interval.

auto<-read.csv("week6/Lecture12/autoanalyser.csv")
# names(auto)
auto.lm<-lm(autoanalyser~true,data=auto)
summary(auto.lm)
## 
## Call:
## lm(formula = autoanalyser ~ true, data = auto)
## 
## Residuals:
##      Min       1Q   Median       3Q      Max 
## -0.23333 -0.09583 -0.03333  0.10417  0.21667 
## 
## Coefficients:
##             Estimate Std. Error t value Pr(>|t|)    
## (Intercept)  0.68333    0.18993   3.598  0.00487 ** 
## true         0.85000    0.04096  20.752  1.5e-09 ***
## ---
## Signif. codes:  0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1
## 
## Residual standard error: 0.1586 on 10 degrees of freedom
## Multiple R-squared:  0.9773, Adjusted R-squared:  0.975 
## F-statistic: 430.6 on 1 and 10 DF,  p-value: 1.496e-09

We also have \[(\mathbf{X}^T\mathbf{X})^{-1} =\frac{1}{180}\left( \begin{array}{cc} 258 & -54 \\ -54 & 12 \\ \end{array} \right)\]

Model diagnostics

Normally at this point, you would check model assumptions and plot data to make sure your fitted model is adequate

#Load needed libraries
library(ggfortify)
library(gridExtra)

#Plot residuals against fitted values and Normal QQ plot
autoplot(auto.lm, which=c(1,2,4,5))

#Create data frame with model residuals
auto.fit<-data.frame(res=auto.lm$residuals,
                        res1=c(auto.lm$residuals[2:nrow(auto)],NA))

#Plot residuals against the previous residuals (lag 1)
plot1<-ggplot(auto.fit, aes(x=res1,y=res)) + 
  geom_point() + 
  labs(y="Residuals", x="Residuals lag 1", title="Residual Independence")

#Plot data
plot2<-ggplot(auto, aes(x = true, y = autoanalyser)) +
          geom_point(size=3.2, alpha = 0.4, col="blue") +
          ggtitle("Autoanalyser against true value") +
  geom_smooth(method = "lm",fullrange=TRUE, color="black",size=2,se=FALSE)

grid.arrange(plot1,plot2,ncol=2)

From residuals against fitted values, residuals seem randomly scattered around the zero line and so we may assume residuals have constant variance and mean zero. Form normal QQ plot, it seems reasonable to assume residuals are normally distributed. From Cook’s distance and leverage there does not appear to be any potential influential observations or outliers. Also, there seems to reason to think residuals are not independent. Lastly from the scatterplot of data, the fitted line appears to provide a reasonable fit to the data and describe the relationship between autoanalyser value and true value.

Confidence Interval

We require an interval estimate for

\[\alpha+6\beta = \mathbf{b}^T\boldsymbol{\beta}\]

\(\boldsymbol{\beta} = \left( \begin{array}{c} \alpha \\ \beta \\ \end{array} \right)\) and \(\mathbf{b} = \left( \begin{array}{c} 1 \\ 6 \\ \end{array} \right)\)

\[\begin{aligned} \mathbf{b}^T\boldsymbol{\hat{\beta}} &= \hat{\alpha}+6 \times \hat{\beta}\\ &=0.683+0.85 \times 6\\ &=5.783\\ RSS&=0.252\\ n&=12\\\end{aligned}\]

t(10;0.975) = 2.228 \[(\mathbf{X}^T\mathbf{X})^{-1} =\frac{1}{180}\left( \begin{array}{cc} 258 & -54 \\ -54 & 12 \\ \end{array} \right)\]

Interval Estimate for \((\alpha+6\beta)\): (5.61, 5.95)

The population mean autoanalyser concentration (when the true concentration is 6 units) is very likely to lie between 5.61 and 5.95 units

Prediction interval

Construct a 95% prediction interval for the autoanalyser concentration \(y\) when the true value \(x\) is 6 and interpret the interval

\[5.783 \pm 2.228 \sqrt {\frac{0.252}{10}(1+0.233)}\]

\[5.783 \pm 0.393\]

(5.390, 6.176)

A future observation for the autoanalyser concentration is highly likely to lie between 5.39 and 6.18 when the true value is 6.