STAT1003 – Statistical Techniques
Dr. Emi Tanaka
Australian National University
These slides are best viewed on a modern browser like Google Chrome on a desktop or laptop. Some interactive components may require some time to fully load.
Let \(\theta\) be some population parameter and let \(\hat{\theta}\) denote a point estimator of \(\theta\).
Consider three different point estimators for \(\mu\):
Error of the estimate is the difference between the estimated value and the parameter, and is composed of:
The formal definition of the bias of a point estimator \(\hat{\theta}\) of a parameter \(\theta\) is
\[\text{Bias}(\hat{\theta}) = E(\hat{\theta}) - \theta\]
A point estimator is unbiased if \(\text{Bias}(\hat{\theta}) = 0\), i.e. if \(\text{E}(\hat{\theta}) = \theta\)
If the bias is positive, the estimator tends to over-estimate the parameter.
If the bias is negative, the estimator tends to under-estimate the parameter.
We have shown before that \(E(\bar{X}) = \mu\), so sample mean \(\bar{X}\) is an unbiased estimator for \(\mu\).
The sample median, \(\tilde{X}\), would:
If we use the average of the first two observations in the sample as an estimator, this would also be an unbiased estimator because
\[ E\left[\frac{1}{2}(X_{1} + X_{2})\right] = \frac{1}{2}\left[E(X_{1}) + E(X_{2})\right] = \frac{1}{2}(\mu + \mu) = \mu. \]
Sampling error, sometimes called sampling uncertainty, describes how much an estimate will tend to vary from one sample to the next.
It is measured by the variance of an estimator.
If \(\hat{\theta}\) is a point estimator of \(\theta\), the variance of \(\hat{\theta}\) is
\[\text{Var}(\hat{\theta}) = E\left((\hat{\theta} - E(\hat{\theta}))^2\right) = E\left(\hat{\theta}^2\right) - \left(E(\hat{\theta})\right)^2\]
\(\text{Var}(\bar{X}) = \dfrac{\sigma^2}{n}\)
The variance decreases when the sample size increases.
We will get a more efficient estimator for the population mean when the sample size is large.
\(\text{Var}(\tilde{X}) \approx \dfrac{\pi\sigma^2}{2n}\) if the population distribution is normal (proof out of scope).
\(\text{Var}\left[\frac{1}{2}(X_1 + X_2)\right] = \frac{1}{4}\left[\text{Var}(X_1) + \text{Var}(X_2)\right] = \frac{1}{4}\left[\sigma^2 + \sigma^2\right] = \dfrac{\sigma^2}{2}\)
The mean squared error (MSE) of a point estimator takes into account both. It is defined as
\[\text{MSE}(\hat{\theta}) = \text{E}\left[(\hat{\theta} - \theta)^2\right] = \left(\text{E}(\hat{\theta}) - \theta\right)^2 + E\left((\hat{\theta} - E(\hat{\theta}))^2\right) = \text{Bias}(\hat{\theta})^2 + \text{Var}(\hat{\theta})\]
A point estimator is said to be consistent if the MSE of the estimator goes to zero as sample size \(n\) increases:
\[\lim_{n \to \infty} \text{MSE}(\hat{\theta}) = 0\]

STAT1003 – Statistical Techniques