HOME

TheInfoList



OR:

In
probability theory Probability theory is the branch of mathematics concerned with probability. Although there are several different probability interpretations, probability theory treats the concept in a rigorous mathematical manner by expressing it through a set o ...
, the Vysochanskij– Petunin inequality gives a lower bound for the
probability Probability is the branch of mathematics concerning numerical descriptions of how likely an Event (probability theory), event is to occur, or how likely it is that a proposition is true. The probability of an event is a number between 0 and ...
that a
random variable A random variable (also called random quantity, aleatory variable, or stochastic variable) is a mathematical formalization of a quantity or object which depends on random events. It is a mapping or a function from possible outcomes (e.g., the po ...
with finite
variance In probability theory and statistics, variance is the expectation of the squared deviation of a random variable from its population mean or sample mean. Variance is a measure of dispersion, meaning it is a measure of how far a set of numbers ...
lies within a certain number of
standard deviation In statistics, the standard deviation is a measure of the amount of variation or dispersion of a set of values. A low standard deviation indicates that the values tend to be close to the mean (also called the expected value) of the set, while ...
s of the variable's
mean There are several kinds of mean in mathematics, especially in statistics. Each mean serves to summarize a given group of data, often to better understand the overall value (magnitude and sign) of a given data set. For a data set, the ''arithme ...
, or equivalently an upper bound for the probability that it lies further away. The sole restrictions on the
distribution Distribution may refer to: Mathematics *Distribution (mathematics), generalized functions used to formulate solutions of partial differential equations * Probability distribution, the probability of a particular value or value range of a vari ...
are that it be
unimodal In mathematics, unimodality means possessing a unique mode. More generally, unimodality means there is only a single highest value, somehow defined, of some mathematical object. Unimodal probability distribution In statistics, a unimodal pr ...
and have finite
variance In probability theory and statistics, variance is the expectation of the squared deviation of a random variable from its population mean or sample mean. Variance is a measure of dispersion, meaning it is a measure of how far a set of numbers ...
. (This implies that it is a continuous probability distribution except at the
mode Mode ( la, modus meaning "manner, tune, measure, due measure, rhythm, melody") may refer to: Arts and entertainment * '' MO''D''E (magazine)'', a defunct U.S. women's fashion magazine * ''Mode'' magazine, a fictional fashion magazine which is ...
, which may have a non-zero probability.)


Theorem

Let X be a random variable with unimodal distribution, and \alpha\in \mathbb R. If we define \rho=\sqrt then for any r>0, :\begin \operatorname(, X-\alpha, \ge r)\le \begin \frac&r\ge \sqrt\rho \\ \frac-\frac&r\le \sqrt\rho \\ \end. \end


Relation to Gauss's inequality

Taking \alpha equal to a mode of X yields the first case of
Gauss's inequality In probability theory, Gauss's inequality (or the Gauss inequality) gives an upper bound on the probability that a unimodal random variable lies more than any given distance from its mode. Let ''X'' be a unimodal random variable with mode ''m'', a ...
.


Tightness of Bound

Without loss of generality, assume \alpha=0 and \rho=1. * If r<1, the left-hand side can equal one, so the bound is useless. * If r\ge \sqrt, the bound is tight when X=0 with probability 1-\frac and is otherwise distributed uniformly in the interval \left \frac,\frac\right/math>. * If 1\le r\le \sqrt, the bound is tight when X=r with probability \frac-\frac and is otherwise distributed uniformly in the interval \left \frac,r\right/math>.


Specialization to mean and variance

If X has mean \mu and finite, non-zero variance \sigma^2, then taking \alpha=\mu and r=\lambda \sigma gives that for any \lambda > \sqrt = 1.63299..., :\operatorname(\left, X-\mu\\geq \lambda\sigma)\leq\frac.


Proof Sketch

For a relatively elementary proof see.Pukelsheim, F., 1994. The Three Sigma Rule. ''The American Statistician'', 48(2), pp.88-91
/ref> The rough idea behind the proof is that there are two cases: one where the mode of X is close to \alpha compared to r, in which case we can show \operatorname(, X-\alpha, \ge r)\le \frac, and one where the mode of X is far from \alpha compared to r, in which case we can show \operatorname(, X-\alpha, \ge r)\le \frac-\frac. Combining these two cases gives \operatorname(, X-\alpha, \ge r)\le \max\left(\frac,\frac-\frac\right). When \frac=\sqrt, the two cases give the same value.


Properties

The theorem refines
Chebyshev's inequality In probability theory, Chebyshev's inequality (also called the Bienaymé–Chebyshev inequality) guarantees that, for a wide class of probability distributions, no more than a certain fraction of values can be more than a certain distance from th ...
by including the factor of 4/9, made possible by the condition that the distribution be unimodal. It is common, in the construction of
control chart Control charts is a graph used in production control to determine whether quality and manufacturing processes are being controlled under stable conditions. (ISO 7870-1) The hourly status is arranged on the graph, and the occurrence of abnormalit ...
s and other statistical heuristics, to set , corresponding to an upper probability bound of 4/81= 0.04938..., and to construct ''3-sigma'' limits to bound ''nearly all'' (i.e. 95%) of the values of a process output. Without unimodality Chebyshev's inequality would give a looser bound of .


One-sided version

An improved version of the Vysochanskij-Petunin inequality for one-sided tail bounds exists. For a unimodal random variable X with mean \mu and variance \sigma^2 , and r \geq 0, the one-sided Vysochanskij-Petunin inequality holds as follows: :\mathbb(X-\mu\geq r)\leq \begin \dfrac\dfrac & \mboxr^\geq\dfrac\sigma^2,\\ \dfrac\dfrac-\dfrac & \mbox \end The one-sided Vysochanskij-Petunin inequality, as well as the related Cantelli inequality, can for instance be relevant in the financial area, in the sense of "how bad can losses get."


Proof

The proof is very similar to that of
Cantelli's inequality In probability theory, Cantelli's inequality (also called the Chebyshev-Cantelli inequality and the one-sided Chebyshev inequality) is an improved version of Chebyshev's inequality for one-sided tail bounds. The inequality states that, for \lambda > ...
. For any u\ge 0, :\begin \mathbb(X-\mu\geq r)&=\mathbb((X+u)-\mu\geq r+u)\\ &\le \mathbb(, (X+u)-\mu), \geq r+u).\\ \end Then we can apply the Vysochanskij-Petunin inequality. With \rho^2=\mathbb E (X+u)-\mu)^2u^2+\sigma^2, we have: : \begin \mathbb(, (X+u)-\mu), \geq r+u) &\le \begin \frac \frac & r+u\ge \sqrt\rho\\ \frac \frac-\frac & r+u\le \sqrt\rho \end. \end As in the proof of Cantelli's inequality, it can be shown that the minimum of \frac over all u\ge 0 is achieved at u=\sigma^2/r. Plugging in this value of u and simplifying yields the desired inequality.


See also

*
Gauss's inequality In probability theory, Gauss's inequality (or the Gauss inequality) gives an upper bound on the probability that a unimodal random variable lies more than any given distance from its mode. Let ''X'' be a unimodal random variable with mode ''m'', a ...
, a similar result for the distance from the mode rather than the mean *
Rule of three (statistics) In statistical analysis, the rule of three states that if a certain event did not occur in a sample with subjects, the interval from 0 to 3/ is a 95% confidence interval for the rate of occurrences in the population. When is greater than 30, t ...
, a similar result for the Bernoulli distribution


References

*
Report (on cancer diagnosis) by Petunin and others stating theorem in English
{{DEFAULTSORT:Vysochanskij-Petunin inequality Probabilistic inequalities Statistical inequalities