Chapter 18 Interval Estimation
18.1 Introduction
In this section, we will explore the concept of an interval estimate. We have seen how the Method of Moments, Maximum Likelihood Estimation and Least Squares can be used to find point estimates of parameters. These estimates often have desirable properties such as unbiasedness (the average value of the sample estimate of the parameter is the true population parameter) and consistency (the sample estimate of the parameter converges to the true population parameter as the sample size tends to infinity). However, for finite samples our sample estimate of the parameter will rarely be equal to the true population parameter. Therefore we construct interval estimates, which allow us to quantify our (un)certainty about parameters.
We start with an exercise for you to attempt to construct intervals which you believe contain the true answer. This will provide motivation for the construction of confidence intervals.
18.2 Confident?
The board game Confident? gets players to give an interval for a numerical answer.
The player with the shortest interval containing the correct answer wins a point.
Do you think you would be good at this?
Attempt the following four questions: Confident? Questions
After you have attempted the questions, watch Video 26 for the answers. The video includes discussion of steps for constructing intervals and how we begin to construct confidence intervals.
Video 26: Confident?
18.3 Confidence intervals
If we are interested in estimating a given parameter we can find some estimator using some appropriate method, for example Method of Moments, Maximum Likelihood Estimation or Least Squares Estimation. is called a point estimator since the estimate of that we report is one particular point in the parameter space.
For example, when we are interested in estimating the percentage of UK residents who are in favour of the Government’s policies, we can collect a random sample of UK residents and compute the sample proportion of the people in favour of the policies. We then report that the Government has, say, a 54% approval rating.
The difficulty that arises, though, is what does 54% mean? How exact is our estimate? The point estimator does not give us that information. Instead it is helpful to also include information about the variability of the estimate given, and that will depend both upon the true underlying variance of the population and the sampling distribution of the estimator that we use.
We have 2 options:
- Report the value of the estimate and the standard deviation of the estimate, which is often called the standard error of the estimate. For example, the Government has a 54% approval rating with a 2% standard error.
- Construct an interval estimate for the parameter which incorporates both information about the point estimate, its standard error, and the sampling distribution of the estimator. For example, a 95% confidence interval for the Government’s approval rating is 52.4% to 55.6%.
Confidence Interval
Let be a fixed value. A confidence interval for the parameter is an interval constructed from a random sample such that if we were to repeat the experiment a large number of times the interval would contain the true value of in of the cases.
Note that the interval will depend on the value of the estimate and the sampling distribution of the estimator.
Suppose that is a random sample from a normal distribution with mean and known variance . Construct a confidence interval for .
Watch Video 27 for the construction of the confidence interval for .
Video 27: Confidence interval for .
Construction of confidence interval for .
First, we need a point estimator for , the mean of the normal distribution. The Method of Moments estimator and MLE for are both , the sample mean.
Next we determine the sampling distribution of the estimator . Since is a random sample from a normal distribution, it follows thatWe want to find endpoints and such that
Note that and are random values which are determined by the random sample.
Note that there exist an infinite number of confidence intervals for . We would like to chose the one that is best, that is, the one for which the length of the interval is the shortest. This will be the interval which is symmetric around if the distribution of is symmetric.
where satisfies . The symmetry of the normal distribution means that also .
Solving for we get,
Suppose that is a random sample from a normal distribution with mean and unknown variance . Construct a confidence interval for .
Again we use , since is the minimum variance unbiased estimator of .
We know that , soIsolating we get,
We make a few observations about the confidence intervals constructed in Example 18.3.2 and Example 18.3.3
- The sample mean (maximum likelihood estimator/method of moments estimator) is contained in our confidence intervals and in these cases the confidence interval is symmetric about .
- The confidence intervals become smaller with . That is, the more data (information) we have, the smaller the confidence interval becomes. Specifically, the confidence intervals decrease at rate . Therefore if we increase the sample size to we will (approximately) half the length of the confidence interval. The exact impact on the confidence interval will depend on the sample mean, and if the variance is unknown, sample variance.
- As becomes smaller, () and become larger. That is, the width of the confidence interval increases as we increase the level of confidence. We will explore in Session 10: Confidence intervals and hypothesis testing the effect of on the confidence interval.
- The -distribution has fatter tails than the normal distribution and has fatter tails for smaller values of . Mathematically, for any and for any , we have that
18.4 Asymptotic distribution of the MLE
Let be the MLE of . Recall that as . Consequently we can construct an approximate confidence interval for , however since is unknown we will also need to approximate with the observed information
Consequently,This method is extremely useful since it is often quite straightforward to evaluate the MLE and the observed information. Nonetheless it is an approximation, and should only be trusted for large values of , though the quality of the approximation will vary from model to model.
Consider independently. Construct an approximate 95% confidence interval for .
For each of the , some is observed and we have
Thus calculating the likelihood and log-likelihood functions:
which is maximised by
Now,
so that
Hence, an approximate 95% confidence interval for is
Consider independently. Construct an approximate 95% confidence interval for .
As noted in Section 10.3, Example 10.3.7, we showed that and
Hence, , and a 95% confidence interval for is
Note that the confidence interval constructed in Example 18.4.2 coincides with the confidence interval constructed in Example 18.3.2 with . This is because the MLE, , satisfies .