Master the mathematical foundation of probability through random variables. Learn discrete and continuous distributions, joint distributions, independence, and sampling distributions essential for statistical inference.
Understanding the mathematical foundation and classification of random variables
Let (Ω, ℱ, P) be a probability space. A single-valued real function ξ(ω) is called a random variable if: for any Borel set B, ξ⁻¹(B) = {ω : ξ(ω) ∈ B} ∈ ℱ
Random variables map sample points ω from the sample space Ω to real numbers, achieving 'quantification of random experiment results'
A random variable that takes on a finite or countably infinite number of values
Described by distribution sequence (probability mass function): P(ξ = xᵢ) = p(xᵢ)
A random variable whose values fill some interval, with a non-negative integrable probability density function p(x)
Described by probability density function (PDF): F(x) = ∫₋∞ˣ p(t)dt
For continuous random variables: P(a < ξ ≤ b) = P(a ≤ ξ ≤ b) = P(a ≤ ξ < b) = P(a < ξ < b) = ∫ₐᵇ p(x)dx
The fundamental tool for describing probability distributions
For any random variable ξ, F(x) = P(ξ ≤ x) for x ∈ ℝ is called the distribution function of ξ
If a ≤ b, then F(a) ≤ F(b)
The probability never decreases as we move right on the real line
F(-∞) = 0, F(+∞) = 1
Probability approaches 0 at negative infinity and 1 at positive infinity
F(x+0) = F(x) for all x
The function is continuous from the right at every point
Step function with jumps at discrete values
Smooth function with F'(x) = p(x) at continuity points
Essential discrete probability distributions and their applications
p ∈ (0,1) (success probability)
Single trial success/failure experiment
n ∈ ℕ (trials), p ∈ (0,1) (success probability)
Number of successes in n independent Bernoulli trials
λ > 0 (rate parameter)
Rare events occurrence (defects, arrivals, accidents)
p ∈ (0,1) (success probability)
Number of trials until first success
n (sample size), M (success states), N (population size)
Sampling without replacement (defective items, card draws)
Fundamental continuous probability distributions and their properties
a, b ∈ ℝ, a < b
Equal probability over an interval (random timing, rounding errors)
μ ∈ ℝ (mean), σ² > 0 (variance)
Natural phenomena (heights, measurement errors, test scores)
λ > 0 (rate parameter)
Lifetime modeling, waiting times between events
α > 0 (shape), β > 0 (rate)
Sum of independent exponential variables, reliability modeling
n ∈ ℕ (degrees of freedom)
Sum of squares of independent standard normal variables
Joint distributions, independence, and conditional distributions
Joint PMF: P(X = xᵢ, Y = yⱼ) = pᵢⱼ with Σᵢ Σⱼ pᵢⱼ = 1
Marginals: P(X = xᵢ) = Σⱼ pᵢⱼ, P(Y = yⱼ) = Σᵢ pᵢⱼ
Joint PDF: p(x,y) with ∫∫ p(x,y)dxdy = 1
Marginals: p_X(x) = ∫ p(x,y)dy, p_Y(y) = ∫ p(x,y)dx
Random variables X and Y are independent if F(x,y) = F_X(x)F_Y(y) for all x,y
In bivariate normal distribution N(μ₁,μ₂,σ₁²,σ₂²,ρ), X and Y are independent iff ρ = 0
P(Y = yⱼ|X = xᵢ) = pᵢⱼ/pᵢ· (when pᵢ· > 0)
p_{Y|X}(y|x) = p(x,y)/p_X(x) (when p_X(x) > 0)
Essential distributions for statistical inference
If X₁, X₂, ..., Xₙ ~ N(0,1) independently, then Y = Σᵢ₌₁ⁿ Xᵢ² ~ χ²(n)
If X ~ N(0,1), Y ~ χ²(n) independently, then T = X/√(Y/n) ~ t(n)
If X ~ χ²(m), Y ~ χ²(n) independently, then F = (X/m)/(Y/n) ~ F(m,n)
Step-by-step solutions to typical random variable problems
In 10 independent trials with success probability 0.3, find P(X = 4)
Binomial distribution applies to fixed number of independent trials with constant success probability
If X ~ N(100, 16), find P(96 < X < 108)
Standardization allows us to use the standard normal table for any normal distribution
For X ~ Exp(λ), prove that P(X > s+t|X > s) = P(X > t)
Memoryless property means past waiting time doesn't affect future waiting time