The Central Limit Theorem

Theorem (The Central Limit Theorem (CLT))

Let be IID with mean and variance . Let . Then

where . In other words,

The Law of Large Numbers and the CLT both describe what happens to the sample mean as the sample size grows, but they emphasize different aspects of that behavior.

  • The LLN is a statement about stability: averaging many observations washes out randomness and pulls the sample mean toward the population mean .
  • The CLT is a statement about the shape of the remaining randomness: after centering at and scaling by , the error in the sample mean is approximately normal.

In that sense, the LLN explains why the average becomes reliable, while the CLT explains what the leftover uncertainty looks like for a large but finite sample. This is why the LLN is most useful when the main point is that an empirical average should be close to the truth, whereas the CLT is most useful when approximating probabilities or building confidence intervals. Informally, the LLN says averaging makes noise die out, while the CLT says the leftover noise is approximately Gaussian and has size about .

The interpretation is that probability statements about can be approximated using a Normal distribution. Its the probability statements that we are approximating, not the random variable itself.

In addition to , there are several forms of notation to denote that the distribution of is converging to a Normal. They all mean the same thing:

Example

Suppose that the number of errors per computer program has a Poisson distribution with mean 5. We get 125 programs. Let be the number of errors in the programs. We want to approximate . Let and . Then,

One important thing to note is that we rarely know . However, we can estimate from by

Interestingly, if is replaced with , the central limit theorem still holds:

Theorem

Assume the same conditions as the CLT. Then,