MathIsimple

Probability Theory – Problem 45: Prove that defines a metric on the space of random variables

Question

For random variables X, YX,~Y, define d(X, Y)=EXY1+XY.d(X,\ Y)=E{\frac{|X-Y|}{1+|X-Y|}}.

(1) Prove that dd defines a metric on the space of random variables. That is, prove that dd satisfies positive definiteness, symmetry, and the triangle inequality.

(2) Prove that the random variables XnX_{n} converge to XX in probability if and only if limnd(Xn, X)=0\operatorname*{lim}_{n\to\infty}d(X_{n},\ X)=0.

Step-by-step solution

Step 1. First, prove that d(X,Y)d(X, Y) satisfies the definition of a metric. (i) Non-negativity and positive definiteness: Since XY0|X-Y| \ge 0, the integrand XY1+XY0\frac{|X-Y|}{1+|X-Y|} \ge 0, so d(X,Y)0d(X, Y) \ge 0 after taking expectations. If d(X,Y)=0d(X, Y) = 0, then the non-negative random variable XY1+XY\frac{|X-Y|}{1+|X-Y|} has expectation 0, which means it equals 0 almost surely, hence XY=0|X-Y|=0 a.s., i.e., X=YX=Y a.s. (ii) Symmetry: Clearly XY=YX|X-Y| = |Y-X|, so d(X,Y)=E[XY1+XY]=E[YX1+YX]=d(Y,X)d(X, Y) = E\left[\frac{|X-Y|}{1+|X-Y|}\right] = E\left[\frac{|Y-X|}{1+|Y-X|}\right] = d(Y, X). (iii) Triangle inequality: Consider the function f(t)=t1+t=111+tf(t) = \frac{t}{1+t} = 1 - \frac{1}{1+t}. For t0t \ge 0, the derivative f(t)=1(1+t)2>0f'(t) = \frac{1}{(1+t)^2} > 0, so f(t)f(t) is monotonically increasing. For any real numbers x,y,zx, y, z, by the absolute value inequality, xzxy+yz|x-z| \le |x-y| + |y-z|. Let a=xy,b=yza = |x-y|, b = |y-z|, so xza+b|x-z| \le a+b. Using the monotonicity of f(t)f(t) and the inequality a+b1+a+b=a1+a+b+b1+a+ba1+a+b1+b\frac{a+b}{1+a+b} = \frac{a}{1+a+b} + \frac{b}{1+a+b} \le \frac{a}{1+a} + \frac{b}{1+b}, we obtain: XZ1+XZXY+YZ1+XY+YZXY1+XY+YZ1+YZ\frac{|X-Z|}{1+|X-Z|} \le \frac{|X-Y| + |Y-Z|}{1 + |X-Y| + |Y-Z|} \le \frac{|X-Y|}{1+|X-Y|} + \frac{|Y-Z|}{1+|Y-Z|}. Taking expectations on both sides yields d(X,Z)d(X,Y)+d(Y,Z)d(X, Z) \le d(X, Y) + d(Y, Z). In summary, dd defines a metric on the space of random variables (in the sense of almost sure equality).

Step 2. Proof of sufficiency: If limnd(Xn,X)=0\lim_{n\to\infty} d(X_n, X) = 0, then XnPXX_n \xrightarrow{P} X. For any given ϵ>0\epsilon > 0, consider the event {XnXϵ}\{|X_n - X| \ge \epsilon\}. On this event, by the monotonicity of f(t)=t1+tf(t)=\frac{t}{1+t}, we have XnX1+XnXϵ1+ϵ\frac{|X_n - X|}{1+|X_n - X|} \ge \frac{\epsilon}{1+\epsilon}. By a generalization of Chebyshev's inequality or directly using properties of expectation: d(Xn,X)=E[XnX1+XnX]E[XnX1+XnXI{XnXϵ}]ϵ1+ϵP(XnXϵ)d(X_n, X) = E\left[\frac{|X_n - X|}{1+|X_n - X|}\right] \ge E\left[\frac{|X_n - X|}{1+|X_n - X|} I_{\{|X_n - X| \ge \epsilon\}}\right] \ge \frac{\epsilon}{1+\epsilon} P(|X_n - X| \ge \epsilon). Rearranging: P(XnXϵ)1+ϵϵd(Xn,X)P(|X_n - X| \ge \epsilon) \le \frac{1+\epsilon}{\epsilon} d(X_n, X). As nn \to \infty, since d(Xn,X)0d(X_n, X) \to 0, we get P(XnXϵ)0P(|X_n - X| \ge \epsilon) \to 0. That is, XnX_n converges to XX in probability.

Step 3. Proof of necessity: If XnPXX_n \xrightarrow{P} X, then limnd(Xn,X)=0\lim_{n\to\infty} d(X_n, X) = 0. For any given ϵ>0\epsilon > 0, split the expectation into two parts: d(Xn,X)=E[XnX1+XnXI{XnXϵ}]+E[XnX1+XnXI{XnX<ϵ}]d(X_n, X) = E\left[\frac{|X_n - X|}{1+|X_n - X|} I_{\{|X_n - X| \ge \epsilon\}}\right] + E\left[\frac{|X_n - X|}{1+|X_n - X|} I_{\{|X_n - X| < \epsilon\}}\right]. For the first term, the integrand is always at most 1, so the first term is at most 1P(XnXϵ)1 \cdot P(|X_n - X| \ge \epsilon). For the second term, when XnX<ϵ|X_n - X| < \epsilon, XnX1+XnX<ϵ1+ϵ<ϵ\frac{|X_n - X|}{1+|X_n - X|} < \frac{\epsilon}{1+\epsilon} < \epsilon, so the second term is at most ϵ\epsilon. That is, d(Xn,X)P(XnXϵ)+ϵd(X_n, X) \le P(|X_n - X| \ge \epsilon) + \epsilon. Since XnPXX_n \xrightarrow{P} X, as nn \to \infty, P(XnXϵ)0P(|X_n - X| \ge \epsilon) \to 0. Therefore lim supnd(Xn,X)ϵ\limsup_{n\to\infty} d(X_n, X) \le \epsilon. Since ϵ\epsilon is arbitrary, limnd(Xn,X)=0\lim_{n\to\infty} d(X_n, X) = 0.

Final answer

QED.

Marking scheme

The following is the rubric based on the official solution (maximum 7 points).


1. Checkpoints (Total max 7)

Part 1: Prove that dd is a metric (max 2 pts)

  • Positive definiteness and symmetry [additive]
  • State that d(X,Y)0d(X, Y) \ge 0 and d(X,Y)=0    X=Yd(X, Y)=0 \iff X=Y a.s. (equal almost surely), and briefly explain symmetry d(X,Y)=d(Y,X)d(X,Y)=d(Y,X).
  • 1 pt
  • Triangle inequality [additive]
  • Use the monotonicity or subadditivity of the function f(t)=t1+tf(t)=\frac{t}{1+t} (i.e., a+b1+a+ba1+a+b1+b\frac{a+b}{1+a+b} \le \frac{a}{1+a} + \frac{b}{1+b}) to derive d(X,Z)d(X,Y)+d(Y,Z)d(X, Z) \le d(X, Y) + d(Y, Z).
  • *If only the triangle inequality formula is stated without proving the core algebraic inequality, no credit is awarded.*
  • 1 pt

Part 2: Prove the equivalence with convergence in probability (max 5 pts)

  • Sufficiency proof: d(Xn,X)0    XnPXd(X_n, X) \to 0 \implies X_n \xrightarrow{P} X [additive]
  • Use Chebyshev's inequality (or Markov's inequality) to establish the connection between PP and EE.
  • Core logic: derive P(XnXϵ)1+ϵϵd(Xn,X)P(|X_n - X| \ge \epsilon) \le \frac{1+\epsilon}{\epsilon} d(X_n, X), or note that on the event {XnXϵ}\{|X_n - X| \ge \epsilon\} the integrand has the lower bound ϵ1+ϵ\frac{\epsilon}{1+\epsilon}.
  • 2 pts
  • Necessity proof: XnPX    d(Xn,X)0X_n \xrightarrow{P} X \implies d(X_n, X) \to 0
  • Score exactly one chain | take the maximum subtotal among chains; do not add points across chains.
  • `Chain A (Truncation/decomposition method)`
  • Decompose the expectation: Split d(Xn,X)d(X_n, X) into integrals over {XnX<ϵ}\{|X_n - X| < \epsilon\} and {XnXϵ}\{|X_n - X| \ge \epsilon\} (or a similar approach). [1 pt]
  • Bounding and taking limits: Correctly bound both parts (first part ϵ\le \epsilon, second part 1P()\le 1 \cdot P(\dots)), and let nn \to \infty to show the limit is 0. [2 pts]
  • `Chain B (Convergence theorem method)`
  • Transfer of convergence in probability: State that XnPX    Yn=XnX1+XnXP0X_n \xrightarrow{P} X \implies Y_n = \frac{|X_n - X|}{1+|X_n - X|} \xrightarrow{P} 0. [1 pt]
  • Citing a theorem: Invoke the Dominated Convergence Theorem (DCT) (dominated by 1) or the Bounded Convergence Theorem to conclude E[Yn]0E[Y_n] \to 0. [2 pts]

Total (max 7)


2. Zero-credit items

  • Merely copying the metric definition formula or the definition of convergence in probability from the problem, without performing any specific derivation.
  • When proving the triangle inequality, directly asserting that XZXY+YZ|X-Z| \le |X-Y| + |Y-Z| implies the inequality for expectations, without addressing the effect of the denominator 1+1+|\cdot|.
  • In Part 2, merely stating "convergence of expectation implies convergence in probability" or vice versa, without proving it specifically for the nonlinear metric dd.

3. Deductions

  • Ignoring almost sure equality (a.s.): When proving positive definiteness, if it is not stated that d(X,Y)=0    X=Yd(X,Y)=0 \implies X=Y holds only in the "almost sure" sense (or P(X=Y)=1P(X=Y)=1), deduct 1 point.
  • Confusing convergence concepts: In the necessity proof, if convergence is incorrectly assumed to be pointwise or almost sure convergence to directly interchange limits and integrals, without mentioning subsequences or properties of convergence in probability, deduct 1 point (in Chain B, if DCT is used, the version under convergence in probability must be made explicit or the subsequence principle must be invoked; otherwise this is treated as a logical gap).
  • Circular reasoning: Using the conclusion to be proved as a basis in the proof (e.g., directly using the fact that dd is a metric to prove convergence), that part receives 0 points.
Ask AI ✨