MathIsimple

Probability Theory – Problem 57: Prove that defines a metric on random variables, i.e., satisfies positive definiteness, symmetry, and the triangle…

Question

For random variables X, YX,~Y, define d(X, Y)=EXY1+XY.d(X,\ Y)=E{\frac{|X-Y|}{1+|X-Y|}}.

(1) Prove that dd defines a metric on random variables, i.e., dd satisfies positive definiteness, symmetry, and the triangle inequality;

(2) Prove that the random variables XnX_{n} converge in probability to XX if and only if limnd(Xn, X)=0\operatorname*{lim}_{n\to\infty}d(X_{n},\ X)=0.

Step-by-step solution

Step 1. First we prove that d(X,Y)d(X, Y) satisfies the definition of a metric. (i) Non-negativity and positive definiteness: Since XY0|X-Y| \ge 0, the integrand XY1+XY0\frac{|X-Y|}{1+|X-Y|} \ge 0, so d(X,Y)0d(X, Y) \ge 0 after taking expectations. If d(X,Y)=0d(X, Y) = 0, then the expectation of the non-negative random variable XY1+XY\frac{|X-Y|}{1+|X-Y|} equals 0, which implies this random variable is 0 almost surely, hence XY=0|X-Y|=0 a.s., i.e., X=YX=Y a.s. (ii) Symmetry: Clearly XY=YX|X-Y| = |Y-X|, so d(X,Y)=E[XY1+XY]=E[YX1+YX]=d(Y,X)d(X, Y) = E\left[\frac{|X-Y|}{1+|X-Y|}\right] = E\left[\frac{|Y-X|}{1+|Y-X|}\right] = d(Y, X). (iii) Triangle inequality: Consider the function f(t)=t1+t=111+tf(t) = \frac{t}{1+t} = 1 - \frac{1}{1+t}. For t0t \ge 0, the derivative f(t)=1(1+t)2>0f'(t) = \frac{1}{(1+t)^2} > 0, so f(t)f(t) is monotonically increasing. For any real numbers x,y,zx, y, z, the absolute value inequality gives xzxy+yz|x-z| \le |x-y| + |y-z|. Let a=xy,b=yza = |x-y|, b = |y-z|, so xza+b|x-z| \le a+b. Using the monotonicity of f(t)f(t) and the inequality a+b1+a+b=a1+a+b+b1+a+ba1+a+b1+b\frac{a+b}{1+a+b} = \frac{a}{1+a+b} + \frac{b}{1+a+b} \le \frac{a}{1+a} + \frac{b}{1+b}, we obtain: XZ1+XZXY+YZ1+XY+YZXY1+XY+YZ1+YZ\frac{|X-Z|}{1+|X-Z|} \le \frac{|X-Y| + |Y-Z|}{1 + |X-Y| + |Y-Z|} \le \frac{|X-Y|}{1+|X-Y|} + \frac{|Y-Z|}{1+|Y-Z|}. Taking expectations on both sides yields d(X,Z)d(X,Y)+d(Y,Z)d(X, Z) \le d(X, Y) + d(Y, Z). In summary, dd defines a metric on the space of random variables (under almost sure equivalence).

Step 2. Proof of sufficiency: If limnd(Xn,X)=0\lim_{n\to\infty} d(X_n, X) = 0, then XnPXX_n \xrightarrow{P} X. For any given ϵ>0\epsilon > 0, consider the event {XnXϵ}\{|X_n - X| \ge \epsilon\}. On this event, by the monotonicity of f(t)=t1+tf(t)=\frac{t}{1+t}, we have XnX1+XnXϵ1+ϵ\frac{|X_n - X|}{1+|X_n - X|} \ge \frac{\epsilon}{1+\epsilon}. By the generalized Chebyshev inequality or directly using properties of expectation: d(Xn,X)=E[XnX1+XnX]E[XnX1+XnXI{XnXϵ}]ϵ1+ϵP(XnXϵ)d(X_n, X) = E\left[\frac{|X_n - X|}{1+|X_n - X|}\right] \ge E\left[\frac{|X_n - X|}{1+|X_n - X|} I_{\{|X_n - X| \ge \epsilon\}}\right] \ge \frac{\epsilon}{1+\epsilon} P(|X_n - X| \ge \epsilon). Rearranging: P(XnXϵ)1+ϵϵd(Xn,X)P(|X_n - X| \ge \epsilon) \le \frac{1+\epsilon}{\epsilon} d(X_n, X). As nn \to \infty, since d(Xn,X)0d(X_n, X) \to 0, we get P(XnXϵ)0P(|X_n - X| \ge \epsilon) \to 0, i.e., XnX_n converges in probability to XX.

Step 3. Proof of necessity: If XnPXX_n \xrightarrow{P} X, then limnd(Xn,X)=0\lim_{n\to\infty} d(X_n, X) = 0. For any given ϵ>0\epsilon > 0, split the expectation into two parts: d(Xn,X)=E[XnX1+XnXI{XnXϵ}]+E[XnX1+XnXI{XnX<ϵ}]d(X_n, X) = E\left[\frac{|X_n - X|}{1+|X_n - X|} I_{\{|X_n - X| \ge \epsilon\}}\right] + E\left[\frac{|X_n - X|}{1+|X_n - X|} I_{\{|X_n - X| < \epsilon\}}\right]. For the first term, the integrand is always at most 1, so the first term is at most 1P(XnXϵ)1 \cdot P(|X_n - X| \ge \epsilon). For the second term, when XnX<ϵ|X_n - X| < \epsilon, we have XnX1+XnX<ϵ1+ϵ<ϵ\frac{|X_n - X|}{1+|X_n - X|} < \frac{\epsilon}{1+\epsilon} < \epsilon, so the second term is at most ϵ\epsilon. Thus d(Xn,X)P(XnXϵ)+ϵd(X_n, X) \le P(|X_n - X| \ge \epsilon) + \epsilon. Since XnPXX_n \xrightarrow{P} X, as nn \to \infty, P(XnXϵ)0P(|X_n - X| \ge \epsilon) \to 0. Therefore lim supnd(Xn,X)ϵ\limsup_{n\to\infty} d(X_n, X) \le \epsilon. Since ϵ\epsilon is arbitrary, limnd(Xn,X)=0\lim_{n\to\infty} d(X_n, X) = 0.

Final answer

QED.

Marking scheme

The following is the marking scheme based on the official solution (maximum 7 points).


I. Checkpoints (Total max 7)

Part 1: Proving dd is a metric (max 2 pts)

  • Positive definiteness and symmetry [additive]
  • State that d(X,Y)0d(X, Y) \ge 0 and d(X,Y)=0    X=Yd(X, Y)=0 \iff X=Y a.s. (almost sure equality), and briefly justify symmetry d(X,Y)=d(Y,X)d(X,Y)=d(Y,X).
  • 1 pt
  • Triangle inequality [additive]
  • Use the monotonicity or subadditivity of f(t)=t1+tf(t)=\frac{t}{1+t} (i.e., a+b1+a+ba1+a+b1+b\frac{a+b}{1+a+b} \le \frac{a}{1+a} + \frac{b}{1+b}) to derive d(X,Z)d(X,Y)+d(Y,Z)d(X, Z) \le d(X, Y) + d(Y, Z).
  • *If only the triangle inequality formula is stated without proving the core algebraic inequality, no credit is awarded.*
  • 1 pt

Part 2: Proving the equivalence with convergence in probability (max 5 pts)

  • Sufficiency proof: d(Xn,X)0    XnPXd(X_n, X) \to 0 \implies X_n \xrightarrow{P} X [additive]
  • Use the Chebyshev (or Markov) inequality to establish the connection between PP and EE.
  • Core logic: derive P(XnXϵ)1+ϵϵd(Xn,X)P(|X_n - X| \ge \epsilon) \le \frac{1+\epsilon}{\epsilon} d(X_n, X) or identify that on the event {XnXϵ}\{|X_n - X| \ge \epsilon\} the integrand has lower bound ϵ1+ϵ\frac{\epsilon}{1+\epsilon}.
  • 2 pts
  • Necessity proof: XnPX    d(Xn,X)0X_n \xrightarrow{P} X \implies d(X_n, X) \to 0
  • Score exactly one chain | take the maximum subtotal among chains; do not add points across chains.
  • `Chain A (Truncation/decomposition method)`
  • Decompose the expectation: Split d(Xn,X)d(X_n, X) into integrals over {XnX<ϵ}\{|X_n - X| < \epsilon\} and {XnXϵ}\{|X_n - X| \ge \epsilon\} (or similar approach). [1 pt]
  • Bounding and taking limits: Correctly bound both parts (first part ϵ\le \epsilon, second part 1P()\le 1 \cdot P(\dots)), and let nn \to \infty to show the limit is 0. [2 pts]
  • `Chain B (Convergence theorem method)`
  • Transfer of convergence in probability: State that XnPX    Yn=XnX1+XnXP0X_n \xrightarrow{P} X \implies Y_n = \frac{|X_n - X|}{1+|X_n - X|} \xrightarrow{P} 0. [1 pt]
  • Cite theorem: Invoke the Dominated Convergence Theorem (DCT) (dominated by 1) or the Bounded Convergence Theorem to conclude E[Yn]0E[Y_n] \to 0. [2 pts]

Total (max 7)


II. Zero-credit items

  • Merely copying the definition of the metric dd or the definition of convergence in probability without any concrete derivation.
  • In proving the triangle inequality, directly claiming that XZXY+YZ|X-Z| \le |X-Y| + |Y-Z| implies the expectation inequality without addressing the effect of the denominator 1+1+|\cdot|.
  • In Part 2, merely stating "convergence in expectation implies convergence in probability" or vice versa without proving it for the specific nonlinear metric dd.

III. Deductions

  • Ignoring almost sure equivalence (a.s.): In proving positive definiteness, if it is not stated that d(X,Y)=0    X=Yd(X,Y)=0 \implies X=Y holds only in the "almost sure" sense (or P(X=Y)=1P(X=Y)=1), deduct 1 point.
  • Confusing convergence concepts: In the necessity proof, if pointwise or almost sure convergence is incorrectly assumed to directly interchange limits and integrals without mentioning subsequences or properties of convergence in probability, deduct 1 point (in Chain B, if DCT is used, the version for convergence in probability must be explicitly stated or the subsequence principle must be invoked; otherwise it is considered a logical gap).
  • Circular reasoning: If the conclusion to be proved is used as a premise (e.g., using the fact that dd is a metric to prove the convergence property), that part receives 0 points.
Ask AI ✨