Cantelli's Inequality
   HOME

TheInfoList



OR:

In probability theory, Cantelli's inequality (also called the Chebyshev-Cantelli inequality and the one-sided Chebyshev inequality) is an improved version of Chebyshev's inequality for one-sided tail bounds. The inequality states that, for \lambda > 0, : \Pr(X-\mathbb ge\lambda) \le \frac, where :X is a real-valued
random variable A random variable (also called random quantity, aleatory variable, or stochastic variable) is a mathematical formalization of a quantity or object which depends on random events. It is a mapping or a function from possible outcomes (e.g., the po ...
, :\Pr is the
probability measure In mathematics, a probability measure is a real-valued function defined on a set of events in a probability space that satisfies measure properties such as ''countable additivity''. The difference between a probability measure and the more gener ...
, :\mathbb /math> is the
expected value In probability theory, the expected value (also called expectation, expectancy, mathematical expectation, mean, average, or first moment) is a generalization of the weighted average. Informally, the expected value is the arithmetic mean of a l ...
of X, :\sigma^2 is the variance of X. Applying the Cantelli inequality to -X gives a bound on the lower tail, : \Pr(X-\mathbb le -\lambda) \le \frac. While the inequality is often attributed to
Francesco Paolo Cantelli Francesco Paolo Cantelli (20 December 187521 July 1966) was an Italian mathematician. He made contributions to celestial mechanics, probability theory, and actuarial science. Biography Cantelli was born in Palermo. He received his doctorate in ...
who published it in 1928, it originates in Chebyshev's work of 1874.Ghosh, B.K., 2002. Probability inequalities related to Markov's theorem. ''The American Statistician'', 56(3), pp.186-190
/ref> When bounding the event random variable deviates from its mean in only one direction (positive or negative), Cantelli's inequality gives an improvement over Chebyshev's inequality. The Chebyshev inequality has "higher moments versions" and "vector versions", and so does the Cantelli inequality.


Comparison to Chebyshev's inequality

For one-sided tail bounds, Cantelli's inequality is better, since Chebyshev's inequality can only get : \Pr(X - \mathbb \geq \lambda) \leq \Pr(, X-\mathbb \ge\lambda) \le \frac. On the other hand, for two-sided tail bounds, Cantelli's inequality gives : \Pr(, X-\mathbb \ge\lambda) = \Pr(X-\mathbb ge\lambda) + \Pr(X-\mathbb le-\lambda) \le \frac, which is always worse than Chebyshev's inequality (when \lambda \geq \sigma; otherwise, both inequalities bound a probability by a value greater than one, and so are trivial).


Proof

Let X be a real-valued random variable with finite variance \sigma^2 and expectation \mu, and define Y = X - \mathbb /math> (so that \mathbb = 0 and \operatorname(Y) = \sigma^2). Then, for any u\geq 0, we have : \Pr( X-\mathbb geq\lambda) = \Pr( Y \geq \lambda) = \Pr( Y + u \geq \lambda + u) \leq \Pr( (Y + u)^2 \geq (\lambda + u)^2 ) \leq \frac = \frac. the last inequality being a consequence of Markov's inequality. As the above holds for any choice of u\in\mathbb, we can choose to apply it with the value that minimizes the function u \geq 0 \mapsto \frac. By differentiating, this can be seen to be u_\ast = \frac, leading to : \Pr( X-\mathbb \geq\lambda) \leq \frac = \frac if \lambda > 0


Generalizations

Using more moments, various stronger inequalities can be shown. He, Zhang, and Zhang showed when \mathbb 0 and \mathbb ^21: : \Pr(X\ge\lambda) \le 1- (2\sqrt-3)\frac


See also

* Chebyshev's inequality * Paley–Zygmund inequality


References

{{reflist Probabilistic inequalities