MathIsimple
Back to Statistics Hub
Probability Theory

Probability Theory Practice 2

Advanced problems on distributions, characteristic functions, order statistics, and convergence

8 Problems
Suggested: 2 hours

Instructions

  • • Try to solve each problem before viewing the solution
  • • Click "Show Solution" to reveal the answer and detailed explanation
  • • Focus on understanding the problem-solving methodology
1Conditional Probability with Multiple Events
Problem

Given events A, B, C where P(A)=P(B)=P(C)=14P(A) = P(B) = P(C) = \frac{1}{4}, P(AB)=P(BC)=0P(AB) = P(BC) = 0, and P(AC)=18P(AC) = \frac{1}{8}.

Find the probability that at least one of A, B, C occurs.

Answer Summary

Translate the word problem into intersections and unions, then use conditional-probability identities and total-probability bookkeeping carefully.

2Characteristic Function Identification
Problem

Determine which of the following are valid characteristic functions:

  • φ1(t)=cos(t)\varphi_1(t) = \cos(t)
  • φ2(t)=cos2(t)\varphi_2(t) = \cos^2(t)
  • φ3(t)=cos(t)\varphi_3(t) = |\cos(t)|
  • φ4(t)=cos(t)\varphi_4(t) = \cos(|t|)
Answer Summary

Match each characteristic function to the distribution features it encodes, especially symmetry, shifts, and convolution behavior.

3Joint Distribution and Marginals
Problem

Random variables X and Y have joint density:

f(x,y)={ce(2x+y)x>0,y>00otherwisef(x, y) = \begin{cases} ce^{-(2x+y)} & x > 0, y > 0 \\ 0 & \text{otherwise} \end{cases}

(1) Find the constant c.

(2) Find the marginal densities fX(x)f_X(x) and fY(y)f_Y(y).

(3) Are X and Y independent?

(4) Find P(X+Y<1)P(X + Y < 1).

Answer Summary

Start from the joint law, sum or integrate to get marginal distributions, and then check whether factorization shows independence.

4Moment Generating Functions
Problem

Let X have moment generating function MX(t)=112tM_X(t) = \frac{1}{1-2t} for t<12t < \frac{1}{2}.

(1) Identify the distribution of X.

(2) Find E[X]E[X] and Var(X)\text{Var}(X).

(3) If X1,X2,,XnX_1, X_2, \ldots, X_n are i.i.d. copies of X, find the MGF of Sn=X1+X2++XnS_n = X_1 + X_2 + \cdots + X_n.

Answer Summary

Differentiate the MGF at zero to recover moments and use the closed form to identify the distributional family when possible.

5Order Statistics
Problem

Let X1,X2,,XnX_1, X_2, \ldots, X_n be i.i.d. uniform random variables on [0, 1].

Let X(1)X(2)X(n)X_{(1)} \leq X_{(2)} \leq \cdots \leq X_{(n)} be the order statistics.

(1) Find the density of X(k)X_{(k)}, the k-th order statistic.

(2) Find E[X(k)]E[X_{(k)}].

(3) Find the joint density of (X(1),X(n))(X_{(1)}, X_{(n)}).

Answer Summary

Use the standard order-statistic density or CDF formulas to move from the parent distribution to the min, max, or kth order statistic.

6Convergence in Distribution
Problem

Let XnX_n have distribution function:

Fn(x)={0x<01(1xn)n0xn1x>nF_n(x) = \begin{cases} 0 & x < 0 \\ 1 - (1 - \frac{x}{n})^n & 0 \leq x \leq n \\ 1 & x > n \end{cases}

(1) Find the limiting distribution as n → ∞.

(2) Identify the limit distribution.

Answer Summary

Check the limiting CDF or characteristic function and focus on the exact notion of convergence being tested, not stronger notions you do not have.

7Conditional Expectation and Variance
Problem

Let N ~ Poisson(λ), and given N = n, let X1,,XNX_1, \ldots, X_N be i.i.d. with mean μ and variance σ².

Define S=i=1NXiS = \sum_{i=1}^N X_i (with S = 0 if N = 0).

(1) Find E[S].

(2) Find Var(S).

Answer Summary

Condition on the simpler sigma-field first, compute E[X|Y] and Var(X|Y), and then apply the tower rule or law of total variance.

8Transformation of Random Variables
Problem

Let X ~ Uniform(0, 1). Find the distribution of:

(1) Y=ln(X)Y = -\ln(X)

(2) Z=X2Z = X^2

(3) W=X1XW = \frac{X}{1-X}

Answer Summary

Choose the right transformation method: inverse/Jacobian for monotone maps, or split the support when the map is not one-to-one.

Ask AI ✨