Mean Squared Error

Estimator

Definition

The quality of a point estimate is sometimes assessed by the mean squared error, or MSE, defined by

Keep in mind that refers to the expectation with respect to the distribution

that generated the data. It does not mean we are averaging over a distribution for . In other words, is treated as fixed but unknown, while is random because it depends on the random sample . Thus, the MSE is the average squared error we would obtain over repeated samples drawn from the population when the true parameter value is .

More generally, even when the true parameter value is unknown, we can still analyze an estimator theoretically by deriving its bias, variance, or MSE as functions of the unknown parameter and the sample size. Then, once we observe a sample and compute the estimator, those theoretical properties help us interpret how reliable the estimate is under repeated sampling.

Theorem

The MSE can be written as

Proof

Let . Then,

Where (1) is due to being a constant, and (2) being a constant.

A shorter proof can be achieved using that for a random variable , :

Proof

Theorem

If and the standard error as , then is consistent, that is, .

Proof

If and , then by the above theorem, . It follows from convergence in quadratic mean that . The result then follows from the convergence properties that implies that .

Predictor

If a vector of predictions is generated from a sample of data points on all random variables, and is the vector of observed values of the variable being predicted, with being the predicted values, then the within-sample MSE of the predictor is computed as

In matrix notation,

where is , and is a column vector.

This version of MSE is used when the goal is prediction rather than estimation of a population parameter. In that setting, we evaluate how close the predicted responses are to the observed outcomes, whereas the estimator version of MSE evaluates how close an estimator tends to be to a fixed but unknown parameter value.

Examples

Mean

Example

Suppose we have a random sample of size from a population, . The usual estimator for the population mean is the sample mean

which has an expected value equal to the true mean (so it is unbiased), and a mean squared error of

where is the population variance.

Variance

Example

The usual estimator for the variance is the corrected sample variance

This is unbiased (its expected value is ), and its MSE is

where is the fourth central moment of the distribution or population.