HOME

TheInfoList



OR:

In
statistics Statistics (from German language, German: ''wikt:Statistik#German, Statistik'', "description of a State (polity), state, a country") is the discipline that concerns the collection, organization, analysis, interpretation, and presentation of ...
, the ''k''th order statistic of a
statistical sample In statistics, quality assurance, and survey methodology, sampling is the selection of a subset (a statistical sample) of individuals from within a statistical population to estimate characteristics of the whole population. Statisticians attempt ...
is equal to its ''k''th-smallest value. Together with rank statistics, order statistics are among the most fundamental tools in non-parametric statistics and
inference Inferences are steps in reasoning, moving from premises to logical consequences; etymologically, the word '' infer'' means to "carry forward". Inference is theoretically traditionally divided into deduction and induction, a distinction that in ...
. Important special cases of the order statistics are the minimum and maximum value of a sample, and (with some qualifications discussed below) the
sample median Sample or samples may refer to: Base meaning * Sample (statistics), a subset of a population – complete data set * Sample (signal), a digital discrete sample of a continuous analog signal * Sample (material), a specimen or small quantity of so ...
and other sample quantiles. When using probability theory to analyze order statistics of random samples from a continuous distribution, the
cumulative distribution function In probability theory and statistics, the cumulative distribution function (CDF) of a real-valued random variable X, or just distribution function of X, evaluated at x, is the probability that X will take a value less than or equal to x. Ev ...
is used to reduce the analysis to the case of order statistics of the
uniform distribution Uniform distribution may refer to: * Continuous uniform distribution * Discrete uniform distribution * Uniform distribution (ecology) * Equidistributed sequence In mathematics, a sequence (''s''1, ''s''2, ''s''3, ...) of real numbers is said to be ...
.


Notation and examples

For example, suppose that four numbers are observed or recorded, resulting in a sample of size 4. If the sample values are :6, 9, 3, 8, the order statistics would be denoted :x_=3,\ \ x_=6,\ \ x_=8,\ \ x_=9,\, where the subscript enclosed in parentheses indicates the th order statistic of the sample. The first order statistic (or smallest order statistic) is always the minimum of the sample, that is, :X_=\min\ where, following a common convention, we use upper-case letters to refer to random variables, and lower-case letters (as above) to refer to their actual observed values. Similarly, for a sample of size , the th order statistic (or largest order statistic) is the maximum, that is, :X_=\max\. The
sample range In statistics, the range of a set of data is the difference between the largest and smallest values, the result of subtracting the sample maximum and minimum. It is expressed in the same units as the data. In descriptive statistics, range is t ...
is the difference between the maximum and minimum. It is a function of the order statistics: :\ = X_-X_. A similar important statistic in exploratory data analysis that is simply related to the order statistics is the sample interquartile range. The sample median may or may not be an order statistic, since there is a single middle value only when the number of observations is odd. More precisely, if for some integer , then the sample median is X_ and so is an order statistic. On the other hand, when is even, and there are two middle values, X_ and X_, and the sample median is some function of the two (usually the average) and hence not an order statistic. Similar remarks apply to all sample quantiles.


Probabilistic analysis

Given any random variables ''X''1, ''X''2..., ''X''''n'', the order statistics X(1), X(2), ..., X(''n'') are also random variables, defined by sorting the values ( realizations) of ''X''1, ..., ''X''''n'' in increasing order. When the random variables ''X''1, ''X''2..., ''X''''n'' form a
sample Sample or samples may refer to: Base meaning * Sample (statistics), a subset of a population – complete data set * Sample (signal), a digital discrete sample of a continuous analog signal * Sample (material), a specimen or small quantity of s ...
they are independent and identically distributed. This is the case treated below. In general, the random variables ''X''1, ..., ''X''''n'' can arise by sampling from more than one population. Then they are independent, but not necessarily identically distributed, and their
joint probability distribution Given two random variables that are defined on the same probability space, the joint probability distribution is the corresponding probability distribution on all possible pairs of outputs. The joint distribution can just as well be considered ...
is given by the
Bapat–Beg theorem In probability theory, the Bapat–Beg theorem gives the joint probability distribution of order statistics of independent but not necessarily identically distributed random variables in terms of the cumulative distribution functions of the rando ...
. From now on, we will assume that the random variables under consideration are continuous and, where convenient, we will also assume that they have a probability density function (PDF), that is, they are absolutely continuous. The peculiarities of the analysis of distributions assigning mass to points (in particular, discrete distributions) are discussed at the end.


Cumulative distribution function of order statistics

For a random sample as above, with cumulative distribution F_X(x), the order statistics for that sample have cumulative distributions as follows (where ''r'' specifies which order statistic): :F_(x) = \sum_^ \binom nj
F_(x) F, or f, is the sixth letter in the Latin alphabet, used in the modern English alphabet, the alphabets of other western European languages and others worldwide. Its name in English is ''ef'' (pronounced ), and the plural is ''efs''. Hist ...
1 - F_(x) 1 (one, unit, unity) is a number representing a single or the only entity. 1 is also a numerical digit and represents a single unit of counting or measurement. For example, a line segment of ''unit length'' is a line segment of length 1. I ...
the corresponding probability density function may be derived from this result, and is found to be :f_(x) = \frac f_(x)
F_(x) F, or f, is the sixth letter in the Latin alphabet, used in the modern English alphabet, the alphabets of other western European languages and others worldwide. Its name in English is ''ef'' (pronounced ), and the plural is ''efs''. Hist ...
1 - F_(x) 1 (one, unit, unity) is a number representing a single or the only entity. 1 is also a numerical digit and represents a single unit of counting or measurement. For example, a line segment of ''unit length'' is a line segment of length 1. I ...
. Moreover, there are two special cases, which have CDFs that are easy to compute. :F_(x) = \operatorname(\max\ \leq x) =
F_(x) F, or f, is the sixth letter in the Latin alphabet, used in the modern English alphabet, the alphabets of other western European languages and others worldwide. Its name in English is ''ef'' (pronounced ), and the plural is ''efs''. Hist ...
n :F_(x) = \operatorname(\min\ \leq x) = 1-
1 - F_(x) 1 (one, unit, unity) is a number representing a single or the only entity. 1 is also a numerical digit and represents a single unit of counting or measurement. For example, a line segment of ''unit length'' is a line segment of length 1. I ...
n Which can be derived by careful consideration of probabilities.


Probability distributions of order statistics


Order statistics sampled from a uniform distribution

In this section we show that the order statistics of the
uniform distribution Uniform distribution may refer to: * Continuous uniform distribution * Discrete uniform distribution * Uniform distribution (ecology) * Equidistributed sequence In mathematics, a sequence (''s''1, ''s''2, ''s''3, ...) of real numbers is said to be ...
on the unit interval have marginal distributions belonging to the beta distribution family. We also give a simple method to derive the joint distribution of any number of order statistics, and finally translate these results to arbitrary continuous distributions using the cdf. We assume throughout this section that X_1, X_2, \ldots, X_n is a random sample drawn from a continuous distribution with cdf F_X. Denoting U_i=F_X(X_i) we obtain the corresponding random sample U_1,\ldots,U_n from the standard
uniform distribution Uniform distribution may refer to: * Continuous uniform distribution * Discrete uniform distribution * Uniform distribution (ecology) * Equidistributed sequence In mathematics, a sequence (''s''1, ''s''2, ''s''3, ...) of real numbers is said to be ...
. Note that the order statistics also satisfy U_=F_X(X_). The probability density function of the order statistic U_ is equal to. :f_(u)=u^(1-u)^ that is, the ''k''th order statistic of the uniform distribution is a beta-distributed random variable. :U_ \sim \operatorname(k,n+1\mathbfk). The proof of these statements is as follows. For U_ to be between ''u'' and ''u'' + ''du'', it is necessary that exactly ''k'' − 1 elements of the sample are smaller than ''u'', and that at least one is between ''u'' and ''u'' + d''u''. The probability that more than one is in this latter interval is already O(du^2), so we have to calculate the probability that exactly ''k'' − 1, 1 and ''n'' − ''k'' observations fall in the intervals (0,u), (u,u+du) and (u+du,1) respectively. This equals (refer to multinomial distribution for details) :u^\cdot du\cdot(1-u-du)^ and the result follows. The mean of this distribution is ''k'' / (''n'' + 1).


The joint distribution of the order statistics of the uniform distribution

Similarly, for ''i'' < ''j'', the joint probability density function of the two order statistics ''U''(''i'') < ''U''(''j'') can be shown to be :f_(u,v) = n! which is (up to terms of higher order than O(du\,dv)) the probability that ''i'' − 1, 1, ''j'' − 1 − ''i'', 1 and ''n'' − ''j'' sample elements fall in the intervals (0,u), (u,u+du), (u+du,v), (v,v+dv), (v+dv,1) respectively. One reasons in an entirely analogous way to derive the higher-order joint distributions. Perhaps surprisingly, the joint density of the ''n'' order statistics turns out to be ''constant'': :f_(u_,u_,\ldots,u_) = n!. One way to understand this is that the unordered sample does have constant density equal to 1, and that there are ''n''! different permutations of the sample corresponding to the same sequence of order statistics. This is related to the fact that 1/''n''! is the volume of the region 0. It is also related with another particularity of order statistics of uniform random variables: It follows from the
BRS-inequality BRS-inequality is the short name for Bruss-Robertson-Steele inequality. This inequality gives a convenient upper bound for the expected maximum number of non-negative random variables one can sum up without exceeding a given upper bound s > 0. Fo ...
that the maximum expected number of uniform U(0,1] random variables one can choose from a sample of size n with a sum up not exceeding 0 is bounded above by \sqrt , which is thus invariant on the set of all s, n with constant product s n . Using the above formulas, one can derive the distribution of the range of the order statistics, that is the distribution of U_-U_, i.e. maximum minus the minimum. More generally, for n\geq k>j\geq 1, U_-U_ also has a beta distribution: U_-U_\sim \operatorname(k-j, n-(k-j)+1)From these formulas we can derive the covariance between two order statistics:\operatorname(U_,U_)=\fracThe formula follows from noting that \operatorname(U_-U_)=\operatorname(U_) + \operatorname(U_)-2\cdot \operatorname(U_,U_) =\frac+\frac-2\cdot \operatorname(U_,U_)and comparing that with \operatorname(U)=\fracwhere U\sim \operatorname(k-j,n-(k-j)+1), which is the actual distribution of the difference.


Order statistics sampled from an exponential distribution

For X_1, X_2, .., X_n a random sample of size ''n'' from an
exponential distribution In probability theory and statistics, the exponential distribution is the probability distribution of the time between events in a Poisson point process, i.e., a process in which events occur continuously and independently at a constant average ...
with parameter ''λ'', the order statistics ''X''(''i'') for ''i'' = 1,2,3, ..., ''n'' each have distribution ::X_ \stackrel \frac\left( \sum_^i \frac \right) where the ''Z''''j'' are iid standard exponential random variables (i.e. with rate parameter 1). This result was first published by
Alfréd Rényi Alfréd Rényi (20 March 1921 – 1 February 1970) was a Hungarian mathematician known for his work in probability theory, though he also made contributions in combinatorics, graph theory, and number theory. Life Rényi was born in Budapest to ...
.


Order statistics sampled from an Erlang distribution

The Laplace transform of order statistics may be sampled from an Erlang distribution via a path counting method .


The joint distribution of the order statistics of an absolutely continuous distribution

If ''F''''X'' is absolutely continuous, it has a density such that dF_X(x)=f_X(x)\,dx, and we can use the substitutions :u=F_X(x) and :du=f_X(x)\,dx to derive the following probability density functions for the order statistics of a sample of size ''n'' drawn from the distribution of ''X'': :f_(x) =\frac
_X(x) X, or x, is the twenty-fourth and third-to-last letter in the Latin alphabet, used in the modern English alphabet, the alphabets of other western European languages and others worldwide. Its name in English is ''"ex"'' (pronounced ), ...
-F_X(x) f_X(x) :f_(x,y) = \frac
_X(x) X, or x, is the twenty-fourth and third-to-last letter in the Latin alphabet, used in the modern English alphabet, the alphabets of other western European languages and others worldwide. Its name in English is ''"ex"'' (pronounced ), ...
_X(y)-F_X(x) -F_X(y)f_X(x)f_X(y) where x\le y :f_(x_1,\ldots,x_n)=n!f_X(x_1)\cdots f_X(x_n) where x_1\le x_2\le \dots \le x_n.


Application: confidence intervals for quantiles

An interesting question is how well the order statistics perform as estimators of the quantiles of the underlying distribution.


A small-sample-size example

The simplest case to consider is how well the sample median estimates the population median. As an example, consider a random sample of size 6. In that case, the sample median is usually defined as the midpoint of the interval delimited by the 3rd and 4th order statistics. However, we know from the preceding discussion that the probability that this interval actually contains the population median is :(1/2)^ = \approx 31\%. Although the sample median is probably among the best distribution-independent point estimates of the population median, what this example illustrates is that it is not a particularly good one in absolute terms. In this particular case, a better confidence interval for the median is the one delimited by the 2nd and 5th order statistics, which contains the population median with probability :\left +\right1/2)^ = \approx 78\%. With such a small sample size, if one wants at least 95% confidence, one is reduced to saying that the median is between the minimum and the maximum of the 6 observations with probability 31/32 or approximately 97%. Size 6 is, in fact, the smallest sample size such that the interval determined by the minimum and the maximum is at least a 95% confidence interval for the population median.


Large sample sizes

For the uniform distribution, as ''n'' tends to infinity, the ''p''th sample quantile is asymptotically normally distributed, since it is approximated by : U_ \sim AN\left(p,\frac\right). For a general distribution ''F'' with a continuous non-zero density at ''F'' −1(''p''), a similar asymptotic normality applies: : X_ \sim AN\left(F^(p),\frac\right) where ''f'' is the density function, and ''F'' −1 is the
quantile function In probability and statistics, the quantile function, associated with a probability distribution of a random variable, specifies the value of the random variable such that the probability of the variable being less than or equal to that value equ ...
associated with ''F''. One of the first people to mention and prove this result was Frederick Mosteller in his seminal paper in 1946. Further research led in the 1960s to the Bahadur representation which provides information about the errorbounds. An interesting observation can be made in the case where the distribution is symmetric, and the population median equals the population mean. In this case, the sample mean, by the central limit theorem, is also asymptotically normally distributed, but with variance σ2''/n'' instead. This asymptotic analysis suggests that the mean outperforms the median in cases of low kurtosis, and vice versa. For example, the median achieves better confidence intervals for the
Laplace distribution In probability theory and statistics, the Laplace distribution is a continuous probability distribution named after Pierre-Simon Laplace. It is also sometimes called the double exponential distribution, because it can be thought of as two exponen ...
, while the mean performs better for ''X'' that are normally distributed.


Proof

It can be shown that : B(k,n+1-k)\ \stackrel\ \frac, where : X = \sum_^ Z_i, \quad Y = \sum_^ Z_i, with ''Zi'' being independent identically distributed exponential random variables with rate 1. Since ''X''/''n'' and ''Y''/''n'' are asymptotically normally distributed by the CLT, our results follow by application of the delta method.


Application: Non-parametric density estimation

Moments of the distribution for the first order statistic can be used to develop a non-parametric density estimator. Suppose, we want to estimate the density f_ at the point x^*. Consider the random variables Y_i = , X_i - x^*, , which are i.i.d with distribution function g_Y(y) = f_X(y + x^*) + f_X(x^* - y). In particular, f_X(x^*) = \frac. The expected value of the first order statistic Y_ given a sample of N total observations yields, : E(Y_) = \frac + \frac \int_^ Q''(z) \delta_(z) \, dz where Q is the quantile function associated with the distribution g_, and \delta_N(z) = (N+1)(1-z)^N. This equation in combination with a jackknifing technique becomes the basis for the following density estimation algorithm, Input: A sample of N observations. \_^M points of density evaluation. Tuning parameter a \in (0,1) (usually 1/3). Output: \_^M estimated density at the points of evaluation. 1: Set m_N = \operatorname(N^) 2: Set s_N = \frac 3: Create an s_N \times m_N matrix M_ which holds m_N subsets with s_N observations each. 4: Create a vector \hat to hold the density evaluations. 5: for \ell = 1 \to M do 6: for k = 1 \to m_N do 7: Find the nearest distance d_ to the current point x_\ell within the kth subset 8: end for 9: Compute the subset average of distances to x_\ell:d_\ell = \sum_^ \frac 10: Compute the density estimate at x_\ell:\hat_\ell = \frac 11: end for 12: return \hat In contrast to the bandwidth/length based tuning parameters for
histogram A histogram is an approximate representation of the distribution of numerical data. The term was first introduced by Karl Pearson. To construct a histogram, the first step is to " bin" (or "bucket") the range of values—that is, divide the ent ...
and kernel based approaches, the tuning parameter for the order statistic based density estimator is the size of sample subsets. Such an estimator is more robust than histogram and kernel based approaches, for example densities like the Cauchy distribution (which lack finite moments) can be inferred without the need for specialized modifications such as IQR based bandwidths. This is because the first moment of the order statistic always exists if the expected value of the underlying distribution does, but the converse is not necessarily true.


Dealing with discrete variables

Suppose X_1,X_2,\ldots,X_n are i.i.d. random variables from a discrete distribution with cumulative distribution function F(x) and
probability mass function In probability and statistics, a probability mass function is a function that gives the probability that a discrete random variable is exactly equal to some value. Sometimes it is also known as the discrete density function. The probability mass ...
f(x). To find the probabilities of the k^\text order statistics, three values are first needed, namely :p_1=P(Xx)=1-F(x). The cumulative distribution function of the k^\text order statistic can be computed by noting that : \begin P(X_\leq x)& =P(\textk\textx) ,\\ & =P(\textn-k\textx) ,\\ & =\sum_^p_3^j(p_1+p_2)^ . \end Similarly, P(X_ is given by : \begin P(X_< x)& =P(\textk\textx) ,\\ & =P(\textn-k\textx) ,\\ & =\sum_^(p_2+p_3)^j(p_1)^ . \end Note that the probability mass function of X_ is just the difference of these values, that is to say : \begin P(X_=x)&=P(X_\leq x)-P(X_< x) ,\\ &=\sum_^\left(p_3^j(p_1+p_2)^-(p_2+p_3)^j(p_1)^\right) ,\\ &=\sum_^\left((1-F(x))^j(F(x))^-(1-F(x)+f(x))^j(F(x)-f(x))^\right). \end


Computing order statistics

The problem of computing the ''k''th smallest (or largest) element of a list is called the selection problem and is solved by a selection algorithm. Although this problem is difficult for very large lists, sophisticated selection algorithms have been created that can solve this problem in time proportional to the number of elements in the list, even if the list is totally unordered. If the data is stored in certain specialized data structures, this time can be brought down to O(log ''n''). In many applications all order statistics are required, in which case a sorting algorithm can be used and the time taken is O(''n'' log ''n'').


See also

* Rankit * Box plot *
BRS-inequality BRS-inequality is the short name for Bruss-Robertson-Steele inequality. This inequality gives a convenient upper bound for the expected maximum number of non-negative random variables one can sum up without exceeding a given upper bound s > 0. Fo ...
* Concomitant (statistics) *
Fisher–Tippett distribution In probability theory and statistics, the generalized extreme value (GEV) distribution is a family of continuous probability distributions developed within extreme value theory to combine the Gumbel distribution, Gumbel, Fréchet distribution, F ...
*
Bapat–Beg theorem In probability theory, the Bapat–Beg theorem gives the joint probability distribution of order statistics of independent but not necessarily identically distributed random variables in terms of the cumulative distribution functions of the rando ...
for the order statistics of independent but not necessarily identically distributed random variables * Bernstein polynomial * L-estimator – linear combinations of order statistics * Rank-size distribution * Selection algorithm


Examples of order statistics

*
Sample maximum and minimum In statistics, the sample maximum and sample minimum, also called the largest observation and smallest observation, are the values of the greatest and least elements of a sample. They are basic summary statistics, used in descriptive statistics ...
* Quantile * Percentile * Decile *
Quartile In statistics, a quartile is a type of quantile which divides the number of data points into four parts, or ''quarters'', of more-or-less equal size. The data must be ordered from smallest to largest to compute quartiles; as such, quartiles are a ...
*
Median In statistics and probability theory, the median is the value separating the higher half from the lower half of a data sample, a population, or a probability distribution. For a data set, it may be thought of as "the middle" value. The basic fe ...


References


External links

* Retrieved Feb 02,2005 * Retrieved Feb 02,2005 * C++ sourc
Dynamic Order Statistics
{{DEFAULTSORT:Order Statistic Nonparametric statistics Summary statistics Permutations