Mean Squared Error
Estimator
Definition
The quality of a point estimate is sometimes assessed by the mean squared error, or MSE, defined by
Keep in mind that refers to the expectation with respect to the distribution
that generated the data. It does not mean we are averaging over a distribution for . In other words, is treated as fixed but unknown, while is random because it depends on the random sample . Thus, the MSE is the average squared error we would obtain over repeated samples drawn from the population when the true parameter value is .
More generally, even when the true parameter value is unknown, we can still analyze an estimator theoretically by deriving its bias, variance, or MSE as functions of the unknown parameter and the sample size. Then, once we observe a sample and compute the estimator, those theoretical properties help us interpret how reliable the estimate is under repeated sampling.
Theorem
The MSE can be written as
Proof
Let . Then,
Where (1) is due to being a constant, and (2) being a constant.
A shorter proof can be achieved using that for a random variable , :
Proof
Theorem
If and the standard error as , then is consistent, that is, .
Proof
If and , then by the above theorem, . It follows from convergence in quadratic mean that . The result then follows from the convergence properties that implies that .
Predictor
If a vector of predictions is generated from a sample of data points on all random variables, and is the vector of observed values of the variable being predicted, with being the predicted values, then the within-sample MSE of the predictor is computed as
In matrix notation,
where is , and is a column vector.
This version of MSE is used when the goal is prediction rather than estimation of a population parameter. In that setting, we evaluate how close the predicted responses are to the observed outcomes, whereas the estimator version of MSE evaluates how close an estimator tends to be to a fixed but unknown parameter value.
Examples
Mean
Example
Suppose we have a random sample of size from a population, . The usual estimator for the population mean is the sample mean
which has an expected value equal to the true mean (so it is unbiased), and a mean squared error of
where is the population variance.
Variance
Example
The usual estimator for the variance is the corrected sample variance
This is unbiased (its expected value is ), and its MSE is
where is the fourth central moment of the distribution or population.