Distribution functions converge weakly to , denoted , if:
for all continuity points of .
Convergence in distribution is equivalent to convergence of characteristic functions.
Convergence in probability is stronger than convergence in distribution.
Slutsky's Lemma: If and , then , , etc.
The sample mean converges in probability to the population mean.
The sample mean converges almost surely to the population mean if and only if the expectation is finite.
If are i.i.d. with and , then:
The normalized sum converges in distribution to standard normal distribution.
If with , then for large :
This provides normal approximation to binomial probabilities, useful when is large.
If are i.i.d. with and , approximate .
For , approximate using normal approximation.
Rigorous proofs of fundamental limit theorems
Since are i.i.d.:
For any :
Since probabilities are non-negative, this proves convergence in probability.
Let . Then , .
Define . We want to show .
Let be the characteristic function of . By Taylor expansion:
Since are i.i.d., the CF of is:
This is the characteristic function of . By the Lévy continuity theorem, convergence of CFs implies convergence in distribution.
Truncate the variables to reduce them to the bounded case, then show the truncation error vanishes.
Show that for any .
By Borel-Cantelli, this implies , proving almost sure convergence.
For the bounded case with fourth moments, use:
The full proof requires more sophisticated techniques (martingale theory or ergodic theory).
The weak law gives convergence in probability of the sample mean to the population mean. The strong law gives almost sure convergence, which is a stronger statement because it holds with probability 1.
Use the Central Limit Theorem for sums or averages of many independent variables with finite variance. After centering and scaling, the distribution approaches a normal law, which justifies common approximations in inference.
Almost sure convergence implies convergence in probability, and convergence in probability implies convergence in distribution. The reverse implications do not hold in general, except for special cases such as convergence in distribution to a constant.
Slutsky's Lemma is used when one sequence converges in distribution and another converges in probability to a constant. It lets you combine them through sums, products, and ratios while preserving convergence in distribution.
Limit theorems justify sample averages, normal approximations, and asymptotic statistical procedures. They are the mathematical backbone of estimation, confidence intervals, and hypothesis testing.