MathIsimple

Stochastic Processes – Problem 14: Prove that the density function of is where ,

Question

Let X1,X2,,XnX_{1}, X_{2}, \ldots, X_{n} be independent continuous random variables with common density function ff. Denote by X(i)X_{(i)} the ii-th smallest among X1,X2,,XnX_{1}, X_{2}, \ldots, X_{n}.

(a) Prove that the density function of X(i)X_{(i)} is fX(i)(x)=n!(i1)!(ni)!(F(x))i1(F(x))nif(x)f_{X_{(i)}}(x)=\frac{n!}{(i-1)!(n-i)!}(F(x))^{i-1}(\overline{F}(x))^{n-i}f(x) where F(x)=xf(y)dyF(x)=\int_{-\infty}^{x}f(y)\,dy, F(x)=1F(x)\overline{F}(x)=1-F(x).

(b) Show that P(X(i)x)=k=in(nk)[F(x)]k[F(x)]nkP(X_{(i)}\leqslant x)=\sum_{k=i}^{n}\binom{n}{k}[F(x)]^{k}[\overline{F}(x)]^{n-k}.

(c) Using the preceding two parts, prove the following probability identity: k=in(nk)yk(1y)nk=0yn!(i1)!(ni)!xi1(1x)nidx.\sum_{k=i}^{n}\binom{n}{k}y^{k}(1-y)^{n-k}=\int_{0}^{y}\frac{n!}{(i-1)!(n-i)!}x^{i-1}(1-x)^{n-i}\,dx.

(d) Let TiT_{i} denote the arrival time of the ii-th event in a Poisson process N(t),t0N(t),t\geqslant0. Find E[TiN(t)=n]E[T_{i}|N(t)=n] (consider the cases ini\leqslant n and i>ni>n separately).

(e) Compute the conditional density function of T1,T2,,Tn1T_{1},T_{2},\dots,T_{n-1} given Tn=tT_{n}=t.

Step-by-step solution

(a) Step 1. Consider the event x<X(i)x+dxx < X_{(i)} \le x+dx. Among the nn i.i.d. random variables X1,,XnX_1,\dots,X_n, partition them into three groups: i1i-1 falling in (,x](-\infty, x], exactly 1 falling in (x,x+dx](x, x+dx], and nin-i falling in (x+dx,)(x+dx, \infty).

Step 2. Compute the probability of this event. By independence and the multinomial formula: P(A)=n!(i1)!1!(ni)![F(x)]i1[P(x<Xx+dx)]1[1F(x+dx)]niP(A) = \frac{n!}{(i-1)!\,1!\,(n-i)!} [F(x)]^{i-1} [P(x < X \le x+dx)]^{1} [1-F(x+dx)]^{n-i}, where P(x<Xx+dx)=f(x)dx+o(dx)P(x < X \le x+dx) = f(x)dx + o(dx).

Step 3. Take the limit to obtain the density. Divide both sides by dxdx and let dx0+dx \to 0^+. By the definition of the derivative: fX(i)(x)=n!(i1)!(ni)![F(x)]i1[1F(x)]nif(x)f_{X_{(i)}}(x) = \frac{n!}{(i-1)!(n-i)!} [F(x)]^{i-1} [1-F(x)]^{n-i} f(x).

(b) {X(i)x}={at least i observationsx}\{X_{(i)} \le x\} = \{\text{at least } i \text{ observations} \le x\}. Let Y=#{j:Xjx}Y = \#\{j: X_j \le x\}. Then YBinomial(n,F(x))Y \sim \text{Binomial}(n, F(x)), so P(X(i)x)=P(Yi)=k=in(nk)[F(x)]k[1F(x)]nkP(X_{(i)} \le x) = P(Y \ge i) = \sum_{k=i}^{n} \binom{n}{k} [F(x)]^{k} [1-F(x)]^{n-k}.

(c) Step 1. From (a), the CDF is FX(i)(x)=xn!(i1)!(ni)![F(t)]i1[1F(t)]nif(t)dtF_{X_{(i)}}(x) = \int_{-\infty}^{x} \frac{n!}{(i-1)!(n-i)!} [F(t)]^{i-1} [1-F(t)]^{n-i} f(t)\,dt.

Step 2. Substituting u=F(t)u = F(t), du=f(t)dtdu = f(t)\,dt: FX(i)(x)=0F(x)n!(i1)!(ni)!ui1(1u)niduF_{X_{(i)}}(x) = \int_{0}^{F(x)} \frac{n!}{(i-1)!(n-i)!} u^{i-1} (1-u)^{n-i}\,du.

Step 3. Setting y=F(x)y = F(x) and equating with (b) yields the identity.

(d) Step 1. When ini \le n: Given N(t)=nN(t)=n, TiT_i has the same distribution as tU(i)t \cdot U_{(i)}, where U(i)U_{(i)} is the ii-th order statistic of nn i.i.d. Uniform(0,1)\text{Uniform}(0,1) variables. Using the Beta function: E[U(i)]=in+1E[U_{(i)}] = \frac{i}{n+1}, so E[TiN(t)=n]=itn+1E[T_i | N(t)=n] = \frac{it}{n+1}.

Step 2. When i>ni > n: By the memoryless property, TiTn=j=n+1iEjT_i - T_n = \sum_{j=n+1}^{i} E_j where Eji.i.d.Exp(λ)E_j \stackrel{\text{i.i.d.}}{\sim} \text{Exp}(\lambda). Thus E[TiN(t)=n]=t+(in)/λE[T_i | N(t)=n] = t + (i-n)/\lambda.

(e) Step 1. Given Tn=tT_n = t, the first n1n-1 arrival times (T1,,Tn1)(T_1,\dots,T_{n-1}) are distributed as the order statistics of n1n-1 i.i.d. Uniform(0,t)\text{Uniform}(0,t) variables.

Step 2. The joint density of these order statistics is (n1)!tn1\frac{(n-1)!}{t^{n-1}} on 0<t1<<tn1<t0<t_1<\cdots<t_{n-1}<t.

Final answer

(a) QED. (b) k=in(nk)[F(x)]k[F(x)]nk\sum_{k=i}^{n}\binom{n}{k}[F(x)]^{k}[\overline{F}(x)]^{n-k}. (c) QED. (d) itn+1\frac{it}{n+1} for ini \le n; t+inλt + \frac{i-n}{\lambda} for i>ni > n. (e) (n1)!tn1\frac{(n-1)!}{t^{n-1}}, for 0<t1<t2<<tn1<t0 < t_1 < t_2 < \cdots < t_{n-1} < t.

Marking scheme

The following is the rubric for this probability problem.


1. Checkpoints (max 7 pts total)

Notes:

  • Points must be awarded according to the logical chains below.
  • Merely copying formulas or listing known conditions earns no credit.
  • For parts with multiple solution paths (e.g., (a) and (e)), score whichever path yields the highest marks; do not combine across paths.
  • (a) Deriving the order statistic density (1 pt)
  • Core derivation: Use the infinitesimal method (considering P(x<X(i)x+dx)P(x < X_{(i)} \le x+dx) via the multinomial probability) or first derive the CDF and differentiate. Must show intermediate steps; writing only the final formula earns 0. [1 pt]
  • (b) Expressing the CDF as a binomial sum (1 pt)
  • Probability reformulation: Identify that {X(i)x}\{X_{(i)} \le x\} is equivalent to "at least ii observations are x\le x" (or similar counting argument), and correctly write the binomial sum. [1 pt]
  • (c) Proving the integral identity (1 pt)
  • Connecting the two expressions: Use FX(i)(x)=xfX(i)(t)dtF_{X_{(i)}}(x) = \int_{-\infty}^x f_{X_{(i)}}(t)dt, apply the substitution u=F(x)u=F(x), and equate the integral form from (a) with the series form from (b). [1 pt]
  • (d) Conditional expectation for the Poisson process (2 pts)
  • Case ini \le n [additive]: Identify that given N(t)=nN(t)=n, the arrival times T1,,TnT_1,\dots,T_n are distributed as the order statistics of nn i.i.d. Uniform(0,t)\text{Uniform}(0,t) variables, and obtain E[TiN(t)=n]=in+1tE[T_i|N(t)=n] = \frac{i}{n+1}t. [1 pt]
  • Case i>ni > n [additive]: Use the memoryless/renewal property of the Poisson process at time tt to show E[TiN(t)=n]=t+inλE[T_i|N(t)=n] = t + \frac{i-n}{\lambda}. [1 pt]
  • (e) Conditional density function (2 pts)
  • Method identification [1 pt]: (A) State that given Tn=tT_n=t, (T1,,Tn1)(T_1,\dots,T_{n-1}) are distributed as the order statistics of n1n-1 i.i.d. Uniform(0,t)\text{Uniform}(0,t) variables; or (B) write the joint density divided by the marginal density of TnT_n.
  • Final result [1 pt]: Write the correct density (n1)!tn1\frac{(n-1)!}{t^{n-1}} with the domain 0<t1<<tn1<t0 < t_1 < \dots < t_{n-1} < t stated or implied.

2. Zero-credit items

  • In (a), merely copying the formula for fX(i)f_{X_{(i)}} with no derivation.
  • In (c), only verifying a special case (e.g., n=1,i=1n=1, i=1) without a general proof.
  • In (d), giving only the unconditional expectation (e.g., i/λi/\lambda), completely ignoring the condition N(t)=nN(t)=n.
  • In (d), incorrectly assuming TiT_i is an nn-equipartition point on (0,t)(0,t), yielding int\frac{i}{n}t.

3. Deductions

  • Logical error (Part d): In the i>ni > n case, computing nn+1t+inλ\frac{n}{n+1}t + \frac{i-n}{\lambda} (incorrectly mixing the conditional expectation of TnT_n with unconditional increments): that sub-part earns 0.
  • Missing domain (Part e): Writing (n1)!tn1\frac{(n-1)!}{t^{n-1}} without specifying 0<t1<<tn1<t0 < t_1 < \dots < t_{n-1} < t and without mentioning order statistics: -1 pt.
  • Conceptual confusion: In (d) or (e), confusing N(t)=nN(t)=n (count condition) with Tn=tT_n=t (time condition), leading to an incorrect distributional model: that part earns 0.

Total: ______ / 7

Ask AI ✨