Chapter 6: Functions of Random Variables

6.7: Order Statistics

6.7: Order Statistics

Let Y1,Y2,...,YnY_1, Y_2, ..., Y_n denote independent continuous random variables with distribution function F(y)F(y) and density function f(y)f(y). We denote the ordered random variables YiY_i by Y(1),Y(2),...,Y(n)Y_{(1)}, Y_{(2)}, ..., Y_{(n)}, where Y(1)Y(2)Y(n)Y_{(1)} \leq Y_{(2)} \leq \cdots \leq Y_{(n)}. Using this notation,

Y(1)=min(Y1,Y2,...,Yn)Y_{(1)} = \min(Y_1, Y_2, ..., Y_n)

is the minimum of the random variables YiY_i, and

Y(n)=max(Y1,Y2,...,Yn)Y_{(n)} = \max(Y_1, Y_2, ..., Y_n)

is the maximum of the random variables YiY_i. Because Y(n)Y_{(n)} is the maximum of Y1,Y2,...,YnY_1, Y_2, ..., Y_n, the event (Y(n)y)(Y_{(n)} \leq y) will occur if and only if the events (Yiy)(Y_i \leq y) occur for every i=1,2,...,ni = 1, 2, ..., n. That is,

P(Y(n)y)=P(Y1y,Y2y,...,Yny).P(Y_{(n)} \leq y) = P(Y_1 \leq y, Y_2 \leq y, ..., Y_n \leq y).

Because the YiY_is are independent and P(Yiy)=F(y)P(Y_i \leq y) = F(y) for i=1,2,...,ni = 1, 2, ..., n, it follows that the distribution function of Y(n)Y_{(n)} is given by

FY(n)(y)=P(Y(n)y)=P(Y1y)P(Y2y)P(Yny)=[F(y)]n.F_{Y_{(n)}}(y) = P(Y_{(n)} \leq y) = P(Y_1 \leq y) P(Y_2 \leq y) \cdots P(Y_n \leq y) = [F(y)]^n.

Letting g(n)(y)g_{(n)}(y) denote the density function of Y(n)Y_{(n)}, we see that, on taking derivatives of both sides,

g(n)(y)=n[F(y)]n1f(y).g_{(n)}(y) = n[F(y)]^{n - 1} f(y).

The density function for Y(1)Y_{(1)} can be found in a similar manner. The distribution function of Y(1)Y_{(1)} is

FY(1)(y)=P(Y(1)y)=1P(Y(1)>y).F_{Y_{(1)}}(y) = P(Y_{(1)} \leq y) = 1 - P(Y_{(1)} > y).

Because Y(1)Y_{(1)} is the minimum of Y1,Y2,...,YnY_1, Y_2, ..., Y_n, it follows that the event (Y(1)>y)(Y_{(1)} > y) occurs if and only if the events (Yi>y)(Y_i > y) occur for i=1,2,...,ni = 1, 2, ..., n. Because the YiY_is are independent and P(Yi>y)=1F(y)P(Y_i > y) = 1 - F(y) for i=1,2,...,ni = 1, 2, ..., n, we see that

FY(1)=P(Y(1)y)=1P(Y(1)>y)=1P(Y1>y,Y2>y,...,Yn>y)=1[P(Y1>y)P(Y2>y)P(Yn>y)]=1[1F(y)]n.\begin{align*} F_{Y_{(1)}} & = P(Y_{(1)} \leq y) = 1 - P(Y_{(1)} > y)\\ & = 1 - P(Y_1 > y, Y_2 > y, ..., Y_n > y)\\ & = 1 - [P(Y_1 > y) P(Y_2 > y) \cdots P(Y_n > y)]\\ & = 1 - [1 - F(y)]^n. \end{align*}

Thus, if g(1)(y)g_{(1)}(y) denotes the density function of Y(1)Y_{(1)}, differentiation of both sides of the last expression yields

g(1)(y)=n[1F(y)]n1f(y).g_{(1)}(y) = n[1 - F(y)]^{n - 1} f(y).

Let us now consider the case n=2n = 2 and find the joint density function for Y(1)Y_{(1)} and Y(2)Y_{(2)}. The event (Y(1)y1,Y(2)y2)(Y_{(1)} \leq y_1, Y_{(2)} \leq y_2) means that either (Y1y1,Y2y2)(Y_1 \leq y_1, Y_2 \leq y_2) or (Y2y1,Y1y2)(Y_2 \leq y_1, Y_1 \leq y_2). [Notice that Y(1)Y_{(1)} could be either Y1Y_1 or Y2Y_2, whichever is smaller.] Therefore, y1y2y_1 \leq y_2, P(Y(1)y1,Y(2)y2)P(Y_{(1)} \leq y_1, Y_{(2)} \leq y_2) is equal to the probability of the union of the two events (Y1y1,Y2y2)(Y_1 \leq y_1, Y_2 \leq y_2) and (Y2y1,Y1y2)(Y_2 \leq y_1, Y_1 \leq y_2). That is,

P(Y(1)y1,Y(2)y2)=P[(Y1y1,Y2y2)(Y2y1,Y1y2)].P(Y_{(1)} \leq y_1, Y_{(2)} \leq y_2) = P[(Y_1 \leq y_1, Y_2 \leq y_2) \cup (Y_2 \leq y_1, Y_1 \leq y_2)].

Using the additive law of probability and recalling that y1y2y_1 \leq y_2, we see that

P(Y(1)y1,Y(2)y2)=P(Y1y1,Y2y2)+P(Y2y1,Y1y2)P(Y1y1,Y2y1).P(Y_{(1)} \leq y_1, Y_{(2)} \leq y_2) = P(Y_1 \leq y_1, Y_2 \leq y_2) + P(Y_2 \leq y_1, Y_1 \leq y_2) - P(Y_1 \leq y_1, Y_2 \leq y_1).

Because Y1Y_1 and Y2Y_2 are independent and P(Yiw)=F(w)P(Y_i \leq w) = F(w), for i=1,2i = 1, 2, it follows that, for y1y2y_1 \leq y_2,

P(Y(1)y1,Y(2)y2)=F(y1)F(y2)+F(y2)F(y1)F(y1)F(y1)=2F(y1)F(y2)[F(y1)]2.\begin{align*} P(Y_{(1)} \leq y_1, Y_{(2)} \leq y_2) & = F(y_1)F(y_2) + F(y_2)F(y_1) - F(y_1)F(y_1)\\ & = 2F(y_1)F(y_2) - [F(y_1)]^2. \end{align*}

If y1>y2y_1 > y_2 (recall that Y(1)Y(2)Y_{(1)} \leq Y_{(2)}),

P(Y(1)y1,Y(2)y2)=P(Y(1)y2,Y(2)y2)=P(Y1y2,Y2y2)=[F(y)]2.\begin{align*} P(Y_{(1)} \leq y_1, Y_{(2)} \leq y_2) & = P(Y_{(1)} \leq y_2, Y_{(2)} \leq y_2)\\ & = P(Y_1 \leq y_2, Y_2 \leq y_2) = [F(y)]^2. \end{align*}

Summarizing, the joint distribution function of Y(1)Y_{(1)} and Y(2)Y_{(2)} is

FY(1),Y(2)(y1,y2)={2F(y1)F(y2)[F(y1)]2,y1y2,[F(y2)]2,y1>y2.F_{Y_{(1)}, Y_{(2)}}(y_1, y_2) = \begin{cases} 2F(y_1)F(y_2) - [F(y_1)]^2, & y_1 \leq y_2,\\ [F(y_2)]^2, & y_1 > y_2. \end{cases}

Letting g(1)(2)(y1,y2)g_{(1)(2)}(y_1, y_2) denote the joint density of Y(1)Y_{(1)} and Y(2)Y_{(2)}, we see that, on differentiating first with respect to y2y_2 and then with respect to y1y_1,

g(1)(2)(y1,y2)={2f(y1)f(y2),y1y2,0,elsewhere.g_{(1)(2)}(y_1, y_2) = \begin{cases} 2f(y_1)f(y_2), & y_1 \leq y_2,\\ 0, & \text{elsewhere.} \end{cases}

The same method can be used to find the joint density of Y(1),Y(2),...,Y(n)Y_{(1)}, Y_{(2)}, ..., Y_{(n)}, which turns out to be

g(1)(2)(n)(y1,y2,...,yn)={n!f(y1)f(y2)f(yn),y1y2yn,0,elsewhere.g_{(1)(2) \cdots (n)}(y_1, y_2, ..., y_n) = \begin{cases} n! f(y_1) f(y_2) \cdots f(y_n), & y_1 \leq y_2 \leq \cdots \leq y_n,\\ 0, & \text{elsewhere.} \end{cases}

The marginal density function for any of the order statistics can be found from this joint density function, but we will not pursue this matter formally in this text.

Although a rigorous derivation of the density function of the kkth-order statistic (kk an integer, 1<k<n1 < k < n) is somewhat complicated, the resulting density function has an intuitively sensible structure. Once that structure is understood, the density can be written down with little difficulty. Think of the density function of a continuous random variable at a particular point as being proportional to the probability that the variable is "close" to that point. That is, if YY is a continuous random variable with density function f(y)f(y), then

P(yYy+dy)f(y)dy.P(y \leq Y \leq y + dy) \approx f(y) dy.

Now consider the kkth-order statistic, Y(k)Y_{(k)}. If the kkth-largest value is near yky_k, then k1k - 1 of the YYs must be less than yky_k, one of the YYs must be near yky_k, and the remaining nkn - k of the YYs must be larger than yky_k. Recall the multinomial distribution, Section 5.9. In the present case, we have three classes of the values of YY:

Class 1: Ys that have values less than yk need k1.Class 2: Ys that have values near yk need 1.Class 3: Ys that have values larger than yk need nk.\begin{align*} \text{Class 1: } & Y \text{s that have values less than } y_k \text{ need } k - 1.\\ \text{Class 2: } & Y \text{s that have values near } y_k \text{ need } 1.\\ \text{Class 3: } & Y \text{s that have values larger than } y_k \text{ need } n - k. \end{align*}

The probabilities of each of these classes are, respectively, p1=P(Y<yk)=F(yk)p_1 = P(Y < y_k) = F(y_k), p2=P(ykYyk+dyk)f(yk)dykp_2 = P(y_k \leq Y \leq y_k + dy_k) \approx f(y_k) dy_k, and p3=P(Y>yk)=1F(yk)p_3 = P(Y > y_k) = 1 - F(y_k). Using the multinomial probabilities discussed earler, we see that

P(ykY(k)yk+dyk)P[(k1) from class 1, 1 from class 2, (nk) from class 3](nk11nk)p1k1p21p3nkn!(k1)!1!(nk)!{[F(yk)]k1f(yk)dyk[1F(yk)]nk}and g(k)(yk)dykn!(k1)!1!(nk)![F(yk)]k1f(yk)[1F(yk)]nkdyk.\begin{align*} P(y_k \leq Y_{(k)} \leq y_k + dy_k) & \approx P[(k - 1) \text{ from class 1, 1 from class 2, } (n - k) \text{ from class 3}]\\ & \approx \binom{n}{k - 1 \quad 1 \quad n - k} p_1^{k - 1} p_2^1 p_3^{n - k}\\ & \approx \frac{n!}{(k - 1)! \, 1! \, (n - k)!} \{[F(y_k)]^{k - 1} f(y_k) dy_k [1 - F(y_k)]^{n - k} \}\\ \text{and } g_{(k)}(y_k) dy_k & \approx \frac{n!}{(k - 1)! \, 1! \, (n - k)!} [F(y_k)]^{k - 1} f(y_k) [1 - F(y_k)]^{n - k} dy_k. \end{align*}
Theorem 6.5

Let Y1,...,YnY_1, ..., Y_n be independent identically distributed continuous random variables with common distribution function F(y)F(y) and common density function f(y)f(y). If Y(k)Y_{(k)} denotes the kkth-order statistic, then the density function of Y(k)Y_{(k)} is given by

g(k)(yk)=n!(k1)!(nk)![F(yk)]k1[1F(yk)]nkf(yk),<yk<.g_{(k)}(y_k) = \frac{n!}{(k - 1)! \, (n - k)!} [F(y_k)]^{k - 1} [1 - F(y_k)]^{n - k} f(y_k), \quad -\infty < y_k < \infty.

If jj and kk are two integers such that 1j<kn1 \leq j < k \leq n, the joint density of Y(j)Y_{(j)} and Y(k)Y_{(k)} is given by

g(j)(k)(yj,yk)=n!(j1)!(k1j)!(nk)!×[F(yj)]j1[F(yk)F(yj)]k1j[1F(yk)]nk×f(yj)f(yk),<yj<yk<.\begin{align*} g_{(j)(k)}(y_j, y_k) & = \frac{n!}{(j - 1)! \, (k - 1 - j)! \, (n - k)!}\\ & \times [F(y_j)]^{j - 1} [F(y_k) - F(y_j)]^{k - 1 - j} [1 - F(y_k)]^{n - k}\\ & \times f(y_j) f(y_k), \quad -\infty < y_j < y_k < \infty. \end{align*}

The heuristic, intuitive derivation of the joint density function given in Theorem 6.5 is similar to that given earlier for the density of a single order statistic. For yj<yky_j < y_k, the joint density can be interpreted as the probability that the jjth largest observation is close to yjy_j and the kkth largest is close to yky_k. Define five classes of values of YY:

Class 1: Ys that have values less than yj need j1.Class 2: Ys that have values near yj need 1.Class 3: Ys that have values between yj and yk need nk.Class 4: Ys that have values near yk need 1.Class 5: Ys that have values larger than yk need nk.\begin{align*} \text{Class 1: } & Y \text{s that have values less than } y_j \text{ need } j - 1.\\ \text{Class 2: } & Y \text{s that have values near } y_j \text{ need } 1.\\ \text{Class 3: } & Y \text{s that have values between } y_j \text{ and } y_k \text{ need } n - k.\\ \text{Class 4: } & Y \text{s that have values near } y_k \text{ need } 1.\\ \text{Class 5: } & Y \text{s that have values larger than } y_k \text{ need } n - k. \end{align*}

Again, use the multinomial distribution to complete the heuristic argument.

Exercises

6.72

Let Y1Y_1 and Y2Y_2 be independent and uniformly distributed over the interval (0,1)(0, 1). Find

a) the probability density function of U1=min(Y1,Y2)U_1 = \min(Y_1, Y_2).

b) E(U1)E(U_1) and V(U1)V(U_1).

6.73

As in Exercise 6.72, let Y1Y_1 and Y2Y_2 be independent and uniformly distributed over the interval (0,1)(0, 1). Find

a) the probability density function of U2=max(Y1,Y2)U_2 = \max(Y_1, Y_2).

b) E(U2)E(U_2) and V(U2)V(U_2).

6.74

Let Y1,Y2,...,YnY_1, Y_2, ..., Y_n be independent, uniformly distributed random variables on the interval [0,θ][0, \theta]. Find the

a) probability distribution function of Y(n)=max(Y1,Y2,...,Yn)Y_{(n)} = \max(Y_1, Y_2, ..., Y_n).

b) density function of Y(n)Y_{(n)}.

c) mean and variance of Y(n)Y_{(n)}.

6.75

Refer to Exercise 6.74. Suppose that the number of minutes that you need to wait for a bus is uniformly distributed on the interval [0,15][0, 15]. If you take the bus five times, what is the probability that your longest wait is less than 10 minutes?

6.76

Let Y1,Y2,...,YnY_1, Y_2, ..., Y_n be independent, uniformly distributed random variables on the interval [0,θ][0, \theta].

a) Find the density function of Y(k)Y_{(k)}, the kkth-order statistic, where kk is an integer between 1 and nn.

b) Use the result from part (a) to find E(Y(k))E(Y_{(k)}).

c) Find V(Y(k))V(Y_{(k)}).

d) Use the result from part (b) to find E(Y(k)Y(k1))E(Y_{(k)} - Y_{(k−1)}), the mean difference between two successive order statistics. Interpret this result.

6.77

Let Y1,Y2,...,YnY_1, Y_2, ..., Y_n be independent, uniformly distributed random variables on the interval [0,θ][0, \theta].

a) Find the joint density function of Y(j)Y_{(j)} and Y(k)Y_{(k)} where jj and kk are integers 1j<kn1 \leq j < k \leq n.

b) Use the result from part (a) to find Cov(Y(j),Y(k))\text{Cov}(Y_{(j)}, Y_{(k)}) when jj and kk are integers 1j<kn1 \leq j < k \leq n.

c) Use the result from part (b) and Exercise 6.76 to find V(Y(k)Y(j))V(Y_{(k)} - Y_{(j)}), the variance of the difference between two order statistics.

6.78

Refer to Exercise 6.76. If Y1,Y2,...,YnY_1, Y_2, ..., Y_n are independent, uniformly distributed random variables on the interval [0,1][0, 1], show that Y(k)Y_{(k)}, the kkth-order statistic, has a beta density function with α=k\alpha = k and β=nk+1\beta = n - k + 1.

6.79

Refer to Exercise 6.77. If Y1,Y2,...,YnY_1, Y_2, ..., Y_n are independent, uniformly distributed random variables on the interval [0,θ][0, \theta], show that U=Y(1)/Y(n)U = Y_{(1)} / Y_{(n)} and Y(n)Y_{(n)} are independent.

6.80

Let Y1,Y2,...,YnY_1, Y_2, ..., Y_n be independent random variables, each with a beta distribution, with α=β=2\alpha = \beta = 2. Find

a) the probability distribution function of Y(n)=max(Y1,Y2,...,Yn)Y_{(n)} = \max(Y_1, Y_2, ..., Y_n).

b) the density function of Y(n)Y_{(n)}.

c) E(Y(n))E(Y_{(n)}) when n=2n = 2.

Solution:

6.81

Let Y1,Y2,...,YnY_1, Y_2, ..., Y_n be independent, exponentially distributed random variables with mean β\beta.

a) Show that Y(1)=min(Y1,Y2,...,Yn)Y_{(1)} = \min(Y_1, Y_2, ..., Y_n) has an exponential distribution, with mean β/n\beta / n.

b) If n=5n = 5 and β=2\beta = 2, find P(Y(1)3.6)P(Y_{(1)} \leq 3.6).

Solution:

6.82

If YY is a continuous random variable and mm is the median of the distribution, then mm is such that P(Ym)=P(Ym)=1/2P(Y \leq m) = P(Y \geq m) = 1 / 2. If Y1,Y2,...,YnY_1, Y_2, ..., Y_n are independent, exponentially distributed random variables with mean β\beta and median mm, Example 6.17 implies that Y(n)=max(Y1,Y2,...,Yn)Y_{(n)} = \max(Y_1, Y_2, ..., Y_n) does not have an exponential distribution. Use the general form of FY(n)(y)F_{Y_{(n)}}(y) to show that P(Y(n)>m)=1(.5)nP(Y_{(n)} > m) = 1 - (.5)^n.

Solution:

6.83

Refer to Exercise 6.82. If Y1,Y2,...,YnY_1, Y_2, ..., Y_n is a random sample from any continuous distribution with mean mm, what is P(Y(n)>m)P(Y_{(n)} > m)?

Solution:

6.84

Refer to Exercise 6.26. The Weibull density function is given by

f(y)={1αmym1eym/α,y>0,0,elsewhere,f(y) = \begin{cases} \frac{1}{\alpha} my^{m - 1} e^{-y^m / \alpha}, & y > 0,\\ 0, & \text{elsewhere,} \end{cases}

where α\alpha and mm are positive constants. If a random sample of size nn is taken from a Weibull distributed population, find the distribution function and density function for Y(1)=min(Y1,Y2,...,Yn)Y_{(1)} = \min(Y_1, Y_2, ..., Y_n). Does Y(1)Y_{(1)} have a Weibull distribution?

Solution:

6.85

Let Y1Y_1 and Y2Y_2 be independent and uniformly distributed over the interval (0,1)(0, 1). Find P(2Y(1)<Y(2))P(2Y_{(1)} < Y_{(2)}).

Solution:

6.86

Let Y1,Y2,...,YnY_1, Y_2, ..., Y_n be independent, exponentially distributed random variables with mean β\beta. Give the

a) density function for Y(k)Y_{(k)}, the kkth-order statistic, where kk is an integer between 11 and nn.

b) joint density function for Y(j)Y_{(j)} and Y(k)Y_{(k)} where jj and kk are integers 1j<kn1 \leq j < k \leq n.

Solution:

6.87

The opening prices per share Y1Y_1 and Y2Y_2 of two similar stocks are independent random variables, each with a density function given by

f(y)={(1/2)e(1/2)(y4),y4,0,elsewhere.f(y) = \begin{cases} (1 / 2) e^{-(1 / 2) (y - 4)}, & y \geq 4,\\ 0, & \text{elsewhere.} \end{cases}

On a given morning, an investor is going to buy shares of whichever stock is less expensive. Find the

a) probability density function for the price per share that the investor will pay.

b) expected cost per share that the investor will pay.

Solution:

6.88

Suppose that the length of time YY it takes a worker to complete a certain task has the probability density function given by

f(y)={e(yθ),y>θ,0,elsewhere,f(y) = \begin{cases} e^{-(y - \theta)}, & y > \theta,\\ 0, & \text{elsewhere,} \end{cases}

where θ\theta is a positive constant that represents the minimum time until task completion. Let Y1,Y2,...,YnY_1, Y_2, ..., Y_n denote a random sample of completion times from this distribution. Find

a) the density function for Y(1)=min(Y1,Y2,...,Yn)Y_{(1)} = \min(Y_1, Y_2, ..., Y_n).

b) E(Y(1))E(Y_{(1)}).

Solution:

6.89

Let Y1,Y2,...,YnY_1, Y_2, ..., Y_n denote a random sample from the uniform distribution f(y)=1,0y1f(y) = 1, 0 \leq y \leq 1. Find the probability density function for the range R=Y(n)Y(1)R = Y_{(n)} - Y_{(1)}.

Solution:

6.90

Suppose that the number of occurrences of a certain event in time interval (0,t)(0, t) has a Poisson distribution. If we know that nn such events have occurred in (0,t)(0, t), then the actual times, measured from 00, for the occurrences of the event in question form an ordered set of random variables, which we denote by W(1)W(2)W(n)W_{(1)} \leq W_{(2)} \leq \cdots \leq W_{(n)}. [W(i)W_{(i)} actually is the waiting time from 00 until the occurrence of the iith event.] It can be shown that the joint density function for W(1),W(2),...,W(n)W_{(1)}, W_{(2)}, ..., W_{(n)} is given by

f(w1,w2,...,wn)={n!tn,w1w2wn,0,elsewhere.f(w_1, w_2, ..., w_n) = \begin{cases} \frac{n!}{t^n}, & w_1 \leq w_2 \leq \cdots \leq w_n,\\ 0, & \text{elsewhere.} \end{cases}

[This is the density function for an ordered sample of size nn from a uniform distribution on the interval (0,t)(0, t).] Suppose that telephone calls coming into a switchboard follow a Poisson distribution with a mean of ten calls per minute. A slow period of two minutes' duration had only four calls. Find the

a) probability that all four calls came in during the first minute; that is, find P(W(4)1)P(W_{(4)} \leq 1).

b) expected waiting time from the start of the two-minute period until the fourth call.

Solution:

6.91

Suppose that nn electronic components, each having an exponentially distributed length of life with mean θ\theta, are put into operation at the same time. The components operate independently and are observed until rr have failed (rn)(r \leq n). Let WjW_j denote the length of time until the jjth failure, with W1W2WrW_1 \leq W_2 \leq \cdots \leq W_r. Let Tj=WjWj1T_j = W_j - W_{j - 1} for j2j \geq 2 and T1=W1T_1 = W_1. Notice that TjT_j measures the time elapsed between successive failures.

a) Show that TjT_j, for j=1,2,...,rj = 1, 2, ..., r, has an exponential distribution with mean θ/(nj+1)\theta / (n - j + 1).

b) Show that

Ur=j=1rWj+(nr)Wr=j=1r(nj+1)TjU_r = \sum_{j = 1}^r W_j + (n - r)W_r = \sum_{j = 1}^r (n - j + 1) T_j

and hence that E(Ur)=rθE(U_r) = r \theta. [UrU_r is called the total observed life, and we can use Ur/rU_r / r as an approximation to (or "estimator" of) θ\theta.]

Solution: