Distribution functions converge weakly to , denoted , if:
for all continuity points of .
Convergence in distribution is equivalent to convergence of characteristic functions.
Convergence in probability is stronger than convergence in distribution.
Slutsky's Lemma: If and , then , , etc.
The sample mean converges in probability to the population mean.
The sample mean converges almost surely to the population mean if and only if the expectation is finite.
If are i.i.d. with and , then:
The normalized sum converges in distribution to standard normal distribution.
If with , then for large :
This provides normal approximation to binomial probabilities, useful when is large.
If are i.i.d. with and , approximate .
For , approximate using normal approximation.
Rigorous proofs of fundamental limit theorems
Since are i.i.d.:
For any :
Since probabilities are non-negative, this proves convergence in probability.
Let . Then , .
Define . We want to show .
Let be the characteristic function of . By Taylor expansion:
Since are i.i.d., the CF of is:
This is the characteristic function of . By the Lévy continuity theorem, convergence of CFs implies convergence in distribution.
Truncate the variables to reduce them to the bounded case, then show the truncation error vanishes.
Show that for any .
By Borel-Cantelli, this implies , proving almost sure convergence.
For the bounded case with fourth moments, use:
The full proof requires more sophisticated techniques (martingale theory or ergodic theory).
Weak Law states convergence in probability: sample mean converges to population mean with probability approaching 1. Strong Law states almost sure convergence: sample mean converges to population mean with probability 1. Strong Law is stronger and requires finite expectation.
CLT applies when you have a sum of many independent random variables with finite variance. The normalized sum (subtract mean, divide by standard deviation) converges to standard normal. This allows normal approximations for sums, averages, and sample statistics.
Almost sure convergence is strongest, implying convergence in probability, which in turn implies convergence in distribution. The reverse implications are not generally true. However, convergence in distribution to a constant implies convergence in probability to that constant.
Use Slutsky's Lemma when you have one sequence converging in distribution and another converging in probability to a constant. You can then combine them: sums, products, and ratios all converge in distribution to the expected limits. This is essential for asymptotic statistics.
Limit theorems justify statistical methods: LLN justifies using sample means as estimates, CLT enables normal approximations for confidence intervals and hypothesis tests, and convergence theory provides theoretical foundation for asymptotic inference. They are fundamental to all statistical practice.