This book is in Open Review. We want your feedback to make the book better for you and other students. You may annotate some text by selecting it with the cursor and then click the on the pop-up menu. You can also see the annotations of others: click the in the upper right hand corner of the page

3.1 Estimation of the Population Mean

Key Concept 3.1

Estimators and Estimates

Estimators are functions of sample data drawn from an unknown population. Estimates are numeric values computed by estimators based on the sample data. Estimators are random variables because they are functions of random data. Estimates are nonrandom numbers.

Think of some economic variable, for example hourly earnings of college graduates, denoted by Y. Suppose we are interested in μY the mean of Y. In order to exactly calculate μY we would have to interview every working graduate in the economy. We simply cannot do this due to time and cost constraints. However, we can draw a random sample of n i.i.d. observations Y1,,Yn and estimate μY using one of the simplest estimators in the sense of Key Concept 3.1 one can think of, that is,

Y¯=1ni=1nYi,

the sample mean of Y. Then again, we could use an even simpler estimator for μY: the very first observation in the sample, Y1. Is Y1 a good estimator? For now, assume that

Yχ122

which is not too unreasonable as hourly income is non-negative and we expect many hourly earnings to be in a range of 5 to 15. Moreover, it is common for income distributions to be skewed to the right — a property of the χ122 distribution.

# plot the chi_12^2 distribution
curve(dchisq(x, df=12), 
      from = 0, 
      to = 40, 
      ylab = "density", 
      xlab = "hourly earnings in Euro")Hide Source
Hide Plot

We now draw a sample of n=100 observations and take the first observation Y1 as an estimate for μY

# set seed for reproducibility
set.seed(1)

# sample from the chi_12^2 distribution, keep only the first observation
rchisq(n = 100, df = 12)[1]
## [1] 8.257893

The estimate 8.26 is not too far away from μY=12 but it is somewhat intuitive that we could do better: the estimator Y1 discards a lot of information and its variance is the population variance:

Var(Y1)=Var(Y)=212=24

This brings us to the following question: What is a good estimator of an unknown parameter in the first place? This question is tackled in Key Concepts 3.2 and 3.3.

Key Concept 3.2

Bias, Consistency and Efficiency

Desirable characteristics of an estimator include unbiasedness, consistency and efficiency.

Unbiasedness:
If the mean of the sampling distribution of some estimator μ^Y for the population mean μY equals μY, E(μ^Y)=μY, the estimator is unbiased for μY. The bias of μ^Y then is 0:

E(μ^Y)μY=0

Consistency:

We want the uncertainty of the estimator μY to decrease as the number of observations in the sample grows. More precisely, we want the probability that the estimate μ^Y falls within a small interval around the true value μY to get increasingly closer to 1 as n grows. We write this as

μ^YpμY.

Variance and efficiency:

We want the estimator to be efficient. Suppose we have two estimators, μ^Y and μY and for some given sample size n it holds that

E(μ^Y)=E(μY)=μY but Var(μ^Y)<Var(μY).

We then prefer to use μ^Y as it has a lower variance than μY, meaning that μ^Y is more efficient in using the information provided by the observations in the sample.