6.7: Order Statistics
Let Y1,Y2,...,Yn denote independent continuous random variables with distribution function F(y) and density function f(y).
We denote the ordered random variables Yi by Y(1),Y(2),...,Y(n), where Y(1)≤Y(2)≤⋯≤Y(n).
Using this notation,
Y(1)=min(Y1,Y2,...,Yn)
is the minimum of the random variables Yi, and
Y(n)=max(Y1,Y2,...,Yn)
is the maximum of the random variables Yi.
Because Y(n) is the maximum of Y1,Y2,...,Yn, the event (Y(n)≤y) will occur if and only if the events (Yi≤y) occur for every i=1,2,...,n.
That is,
P(Y(n)≤y)=P(Y1≤y,Y2≤y,...,Yn≤y).
Because the Yis are independent and P(Yi≤y)=F(y) for i=1,2,...,n, it follows that the distribution function of Y(n) is given by
FY(n)(y)=P(Y(n)≤y)=P(Y1≤y)P(Y2≤y)⋯P(Yn≤y)=[F(y)]n.
Letting g(n)(y) denote the density function of Y(n), we see that, on taking derivatives of both sides,
g(n)(y)=n[F(y)]n−1f(y).
The density function for Y(1) can be found in a similar manner.
The distribution function of Y(1) is
FY(1)(y)=P(Y(1)≤y)=1−P(Y(1)>y).
Because Y(1) is the minimum of Y1,Y2,...,Yn, it follows that the event (Y(1)>y) occurs if and only if the events (Yi>y) occur for i=1,2,...,n.
Because the Yis are independent and P(Yi>y)=1−F(y) for i=1,2,...,n, we see that
FY(1)=P(Y(1)≤y)=1−P(Y(1)>y)=1−P(Y1>y,Y2>y,...,Yn>y)=1−[P(Y1>y)P(Y2>y)⋯P(Yn>y)]=1−[1−F(y)]n.
Thus, if g(1)(y) denotes the density function of Y(1), differentiation of both sides of the last expression yields
g(1)(y)=n[1−F(y)]n−1f(y).
Let us now consider the case n=2 and find the joint density function for Y(1) and Y(2).
The event (Y(1)≤y1,Y(2)≤y2) means that either (Y1≤y1,Y2≤y2) or (Y2≤y1,Y1≤y2).
[Notice that Y(1) could be either Y1 or Y2, whichever is smaller.]
Therefore, y1≤y2, P(Y(1)≤y1,Y(2)≤y2) is equal to the probability of the union of the two events (Y1≤y1,Y2≤y2) and (Y2≤y1,Y1≤y2).
That is,
P(Y(1)≤y1,Y(2)≤y2)=P[(Y1≤y1,Y2≤y2)∪(Y2≤y1,Y1≤y2)].
Using the additive law of probability and recalling that y1≤y2, we see that
P(Y(1)≤y1,Y(2)≤y2)=P(Y1≤y1,Y2≤y2)+P(Y2≤y1,Y1≤y2)−P(Y1≤y1,Y2≤y1).
Because Y1 and Y2 are independent and P(Yi≤w)=F(w), for i=1,2, it follows that, for y1≤y2,
P(Y(1)≤y1,Y(2)≤y2)=F(y1)F(y2)+F(y2)F(y1)−F(y1)F(y1)=2F(y1)F(y2)−[F(y1)]2.
If y1>y2 (recall that Y(1)≤Y(2)),
P(Y(1)≤y1,Y(2)≤y2)=P(Y(1)≤y2,Y(2)≤y2)=P(Y1≤y2,Y2≤y2)=[F(y)]2.
Summarizing, the joint distribution function of Y(1) and Y(2) is
FY(1),Y(2)(y1,y2)={2F(y1)F(y2)−[F(y1)]2,[F(y2)]2,y1≤y2,y1>y2.
Letting g(1)(2)(y1,y2) denote the joint density of Y(1) and Y(2), we see that, on differentiating first with respect to y2 and then with respect to y1,
g(1)(2)(y1,y2)={2f(y1)f(y2),0,y1≤y2,elsewhere.
The same method can be used to find the joint density of Y(1),Y(2),...,Y(n), which turns out to be
g(1)(2)⋯(n)(y1,y2,...,yn)={n!f(y1)f(y2)⋯f(yn),0,y1≤y2≤⋯≤yn,elsewhere.
The marginal density function for any of the order statistics can be found from this joint density function, but we will not pursue this matter formally in this text.
Although a rigorous derivation of the density function of the kth-order statistic (k an integer, 1<k<n) is somewhat complicated, the resulting density function has an intuitively sensible structure.
Once that structure is understood, the density can be written down with little difficulty.
Think of the density function of a continuous random variable at a particular point as being proportional to the probability that the variable is "close" to that point.
That is, if Y is a continuous random variable with density function f(y), then
P(y≤Y≤y+dy)≈f(y)dy.
Now consider the kth-order statistic, Y(k).
If the kth-largest value is near yk, then k−1 of the Ys must be less than yk, one of the Ys must be near yk, and the remaining n−k of the Ys must be larger than yk.
Recall the multinomial distribution, Section 5.9.
In the present case, we have three classes of the values of Y:
Class 1: Class 2: Class 3: Ys that have values less than yk need k−1.Ys that have values near yk need 1.Ys that have values larger than yk need n−k.
The probabilities of each of these classes are, respectively, p1=P(Y<yk)=F(yk), p2=P(yk≤Y≤yk+dyk)≈f(yk)dyk, and p3=P(Y>yk)=1−F(yk).
Using the multinomial probabilities discussed earler, we see that
P(yk≤Y(k)≤yk+dyk)and g(k)(yk)dyk≈P[(k−1) from class 1, 1 from class 2, (n−k) from class 3]≈(k−11n−kn)p1k−1p21p3n−k≈(k−1)!1!(n−k)!n!{[F(yk)]k−1f(yk)dyk[1−F(yk)]n−k}≈(k−1)!1!(n−k)!n![F(yk)]k−1f(yk)[1−F(yk)]n−kdyk.
Let Y1,...,Yn be independent identically distributed continuous random variables with common distribution function F(y) and common density function f(y).
If Y(k) denotes the kth-order statistic, then the density function of Y(k) is given by
g(k)(yk)=(k−1)!(n−k)!n![F(yk)]k−1[1−F(yk)]n−kf(yk),−∞<yk<∞.If j and k are two integers such that 1≤j<k≤n, the joint density of Y(j) and Y(k) is given by
g(j)(k)(yj,yk)=(j−1)!(k−1−j)!(n−k)!n!×[F(yj)]j−1[F(yk)−F(yj)]k−1−j[1−F(yk)]n−k×f(yj)f(yk),−∞<yj<yk<∞.
The heuristic, intuitive derivation of the joint density function given in Theorem 6.5 is similar to that given earlier for the density of a single order statistic.
For yj<yk, the joint density can be interpreted as the probability that the jth largest observation is close to yj and the kth largest is close to yk.
Define five classes of values of Y:
Class 1: Class 2: Class 3: Class 4: Class 5: Ys that have values less than yj need j−1.Ys that have values near yj need 1.Ys that have values between yj and yk need n−k.Ys that have values near yk need 1.Ys that have values larger than yk need n−k.
Again, use the multinomial distribution to complete the heuristic argument.
Let Y1 and Y2 be independent and uniformly distributed over the interval (0,1).
Find
a) the probability density function of U1=min(Y1,Y2).
b) E(U1) and V(U1).
As in Exercise 6.72, let Y1 and Y2 be independent and uniformly distributed over the interval (0,1).
Find
a) the probability density function of U2=max(Y1,Y2).
b) E(U2) and V(U2).
Let Y1,Y2,...,Yn be independent, uniformly distributed random variables on the interval [0,θ].
Find the
a) probability distribution function of Y(n)=max(Y1,Y2,...,Yn).
b) density function of Y(n).
c) mean and variance of Y(n).
Refer to Exercise 6.74.
Suppose that the number of minutes that you need to wait for a bus is uniformly distributed on the interval [0,15].
If you take the bus five times, what is the probability that your longest wait is less than 10 minutes?
Let Y1,Y2,...,Yn be independent, uniformly distributed random variables on the interval [0,θ].
a) Find the density function of Y(k), the kth-order statistic, where k is an integer between 1 and n.
b) Use the result from part (a) to find E(Y(k)).
c) Find V(Y(k)).
d) Use the result from part (b) to find E(Y(k)−Y(k−1)), the mean difference between two successive order statistics.
Interpret this result.
Let Y1,Y2,...,Yn be independent, uniformly distributed random variables on the interval [0,θ].
a) Find the joint density function of Y(j) and Y(k) where j and k are integers 1≤j<k≤n.
b) Use the result from part (a) to find Cov(Y(j),Y(k)) when j and k are integers 1≤j<k≤n.
c) Use the result from part (b) and Exercise 6.76 to find V(Y(k)−Y(j)), the variance of the difference between two order statistics.
Refer to Exercise 6.76.
If Y1,Y2,...,Yn are independent, uniformly distributed random variables on the interval [0,1], show that Y(k), the kth-order statistic, has a beta density function with α=k and β=n−k+1.
Refer to Exercise 6.77.
If Y1,Y2,...,Yn are independent, uniformly distributed random variables on the interval [0,θ], show that U=Y(1)/Y(n) and Y(n) are independent.
Let Y1,Y2,...,Yn be independent random variables, each with a beta distribution, with α=β=2.
Find
a) the probability distribution function of Y(n)=max(Y1,Y2,...,Yn).
b) the density function of Y(n).
c) E(Y(n)) when n=2.
Solution:
Let Y1,Y2,...,Yn be independent, exponentially distributed random variables with mean β.
a) Show that Y(1)=min(Y1,Y2,...,Yn) has an exponential distribution, with mean β/n.
b) If n=5 and β=2, find P(Y(1)≤3.6).
Solution:
If Y is a continuous random variable and m is the median of the distribution, then m is such that P(Y≤m)=P(Y≥m)=1/2.
If Y1,Y2,...,Yn are independent, exponentially distributed random variables with mean β and median m, Example 6.17 implies that Y(n)=max(Y1,Y2,...,Yn) does not have an exponential distribution.
Use the general form of FY(n)(y) to show that P(Y(n)>m)=1−(.5)n.
Solution:
Refer to Exercise 6.82.
If Y1,Y2,...,Yn is a random sample from any continuous distribution with mean m, what is P(Y(n)>m)?
Solution:
Refer to Exercise 6.26.
The Weibull density function is given by
f(y)={α1mym−1e−ym/α,0,y>0,elsewhere,
where α and m are positive constants.
If a random sample of size n is taken from a Weibull distributed population, find the distribution function and density function for Y(1)=min(Y1,Y2,...,Yn).
Does Y(1) have a Weibull distribution?
Solution:
Let Y1 and Y2 be independent and uniformly distributed over the interval (0,1).
Find P(2Y(1)<Y(2)).
Solution:
Let Y1,Y2,...,Yn be independent, exponentially distributed random variables with mean β.
Give the
a) density function for Y(k), the kth-order statistic, where k is an integer between 1 and n.
b) joint density function for Y(j) and Y(k) where j and k are integers 1≤j<k≤n.
Solution:
The opening prices per share Y1 and Y2 of two similar stocks are independent random variables, each with a density function given by
f(y)={(1/2)e−(1/2)(y−4),0,y≥4,elsewhere.
On a given morning, an investor is going to buy shares of whichever stock is less expensive.
Find the
a) probability density function for the price per share that the investor will pay.
b) expected cost per share that the investor will pay.
Solution:
Suppose that the length of time Y it takes a worker to complete a certain task has the probability density function given by
f(y)={e−(y−θ),0,y>θ,elsewhere,
where θ is a positive constant that represents the minimum time until task completion.
Let Y1,Y2,...,Yn denote a random sample of completion times from this distribution.
Find
a) the density function for Y(1)=min(Y1,Y2,...,Yn).
b) E(Y(1)).
Solution:
Let Y1,Y2,...,Yn denote a random sample from the uniform distribution f(y)=1,0≤y≤1.
Find the probability density function for the range R=Y(n)−Y(1).
Solution:
Suppose that the number of occurrences of a certain event in time interval (0,t) has a Poisson distribution.
If we know that n such events have occurred in (0,t), then the actual times, measured from 0, for the occurrences of the event in question form an ordered set of random variables, which we denote by W(1)≤W(2)≤⋯≤W(n).
[W(i) actually is the waiting time from 0 until the occurrence of the ith event.]
It can be shown that the joint density function for W(1),W(2),...,W(n) is given by
f(w1,w2,...,wn)={tnn!,0,w1≤w2≤⋯≤wn,elsewhere.
[This is the density function for an ordered sample of size n from a uniform distribution on the interval (0,t).]
Suppose that telephone calls coming into a switchboard follow a Poisson distribution with a mean of ten calls per minute.
A slow period of two minutes' duration had only four calls.
Find the
a) probability that all four calls came in during the first minute; that is, find P(W(4)≤1).
b) expected waiting time from the start of the two-minute period until the fourth call.
Solution:
Suppose that n electronic components, each having an exponentially distributed length of life with mean θ, are put into operation at the same time.
The components operate independently and are observed until r have failed (r≤n).
Let Wj denote the length of time until the jth failure, with W1≤W2≤⋯≤Wr.
Let Tj=Wj−Wj−1 for j≥2 and T1=W1.
Notice that Tj measures the time elapsed between successive failures.
a) Show that Tj, for j=1,2,...,r, has an exponential distribution with mean θ/(n−j+1).
b) Show that
Ur=j=1∑rWj+(n−r)Wr=j=1∑r(n−j+1)Tj
and hence that E(Ur)=rθ.
[Ur is called the total observed life, and we can use Ur/r as an approximation to (or "estimator" of) θ.]
Solution: