Bernstein Inequalities (probability Theory)
   HOME

TheInfoList



OR:

In
probability theory Probability theory is the branch of mathematics concerned with probability. Although there are several different probability interpretations, probability theory treats the concept in a rigorous mathematical manner by expressing it through a set o ...
, Bernstein inequalities give bounds on the probability that the sum of random variables deviates from its mean. In the simplest case, let ''X''1, ..., ''X''''n'' be independent Bernoulli random variables taking values +1 and −1 with probability 1/2 (this distribution is also known as the
Rademacher distribution In probability theory and statistics, the Rademacher distribution (which is named after Hans Rademacher) is a discrete probability distribution where a random variate ''X'' has a 50% chance of being +1 and a 50% chance of being -1. A series ( ...
), then for every positive \varepsilon, :\mathbb\left (\left, \frac\sum_^n X_i\ > \varepsilon \right ) \leq 2\exp \left (-\frac \right). Bernstein inequalities were proved and published by
Sergei Bernstein Sergei Natanovich Bernstein (russian: Серге́й Ната́нович Бернште́йн, sometimes Romanized as ; 5 March 1880 – 26 October 1968) was a Ukrainian and Russian mathematician of Jewish origin known for contributions to parti ...
in the 1920s and 1930s.J.V.Uspensky, "Introduction to Mathematical Probability", McGraw-Hill Book Company, 1937 Later, these inequalities were rediscovered several times in various forms. Thus, special cases of the Bernstein inequalities are also known as the
Chernoff bound In probability theory, the Chernoff bound gives exponentially decreasing bounds on tail distributions of sums of independent random variables. Despite being named after Herman Chernoff, the author of the paper it first appeared in, the result is d ...
,
Hoeffding's inequality In probability theory, Hoeffding's inequality provides an upper bound on the probability that the sum of bounded independent random variables deviates from its expected value by more than a certain amount. Hoeffding's inequality was proven by Wassi ...
and
Azuma's inequality In probability theory, the Azuma–Hoeffding inequality (named after Kazuoki Azuma and Wassily Hoeffding) gives a concentration inequality, concentration result for the values of martingale (probability theory), martingales that have bounded differe ...
.


Some of the inequalities

1. Let X_1, \ldots, X_n be independent zero-mean random variables. Suppose that , X_i, \leq M almost surely, for all i. Then, for all positive t, :\mathbb \left (\sum_^n X_i \geq t \right ) \leq \exp \left ( -\frac \right ). 2. Let X_1, \ldots, X_n be independent zero-mean random variables. Suppose that for some positive real L and every integer k \geq 2, : \mathbb \left X_i^k \right , \right \leq \frac \mathbb \left _i^2\rightL^ k! Then :\mathbb \left (\sum_^n X_i \geq 2t \sqrt \right ) < \exp(-t^2), \qquad \text\quad 0 \leq t \leq \frac\sqrt. 3. Let X_1, \ldots, X_n be independent zero-mean random variables. Suppose that : \mathbb \left X_i^k \right , \right \leq \frac \left(\frac\right)^ for all integer k \geq 4. Denote : A_k = \sum \mathbb \left X_i^k\right Then, : \mathbb \left( \left, \sum_^n X_j - \frac \\geq \sqrt \, t \left 1 + \frac \right\right) < 2 \exp (- t^2), \qquad \text \quad 0 < t \leq \frac. 4. Bernstein also proved generalizations of the inequalities above to weakly dependent random variables. For example, inequality (2) can be extended as follows. Let X_1, \ldots, X_n be possibly non-independent random variables. Suppose that for all integers i>0, :\begin \mathbb \left. \left X_1, \ldots, X_ \right &= 0, \\ \mathbb \left. \left X_1, \ldots, X_ \right &\leq R_i \mathbb \left X_i^2 \right \\ \mathbb \left. \left X_1, \ldots, X_ \right &\leq \tfrac \mathbb \left. \left X_1, \ldots, X_ \right L^ k! \end Then :\mathbb \left( \sum_^n X_i \geq 2t \sqrt \right) < \exp(-t^2), \qquad \text\quad 0 < t \leq \frac \sqrt. More general results for martingales can be found in Fan et al. (2015).


Proofs

The proofs are based on an application of
Markov's inequality In probability theory, Markov's inequality gives an upper bound for the probability that a non-negative function (mathematics), function of a random variable is greater than or equal to some positive Constant (mathematics), constant. It is named a ...
to the random variable : \exp \left ( \lambda \sum_^n X_j \right ), for a suitable choice of the parameter \lambda > 0.


Generalizations

The Bernstein inequality could be generalized to Gaussian random matrices. Let G = g^H A g + 2 \operatorname(g^H a) be a scalar where A is a complex Hermitian matrix and a is complex vector of size N. The vector g \sim \mathcal(0,I) is a Gaussian vector of size N. Then for any \sigma \geq 0, we have :\mathbb \left( G \leq \operatorname(A) - \sqrt\sqrt - \sigma s^-(A) \right) < \exp(-\sigma), where \operatorname is the
vectorization Vectorization may refer to: Computing * Array programming, a style of computer programming where operations are applied to whole arrays instead of individual elements * Automatic vectorization, a compiler optimization that transforms loops to vec ...
operation. Also, s^- (A) = \max(\lambda_(-A),0) and \lambda_(-A) is the maximum eigenvalue of -A. The proof is detailed in here. Another similar inequality is also formulated as :\mathbb \left( G \geq \operatorname(A) + \sqrt\sqrt + \sigma s^+(A) \right) < \exp(-\sigma), where s^+(A) = \max(\lambda_(A),0) and \lambda_(A) is the maximum eigenvalue of A.


See also

*
Concentration inequality In probability theory, concentration inequalities provide bounds on how a random variable deviates from some value (typically, its expected value). The law of large numbers of classical probability theory states that sums of independent random vari ...
- a summary of tail-bounds on random variables. *
Hoeffding's inequality In probability theory, Hoeffding's inequality provides an upper bound on the probability that the sum of bounded independent random variables deviates from its expected value by more than a certain amount. Hoeffding's inequality was proven by Wassi ...


References

(according to: S.N.Bernstein, Collected Works, Nauka, 1964) A modern translation of some of these results can also be found in {{DEFAULTSORT:Bernstein Inequalities (Probability Theory) Probabilistic inequalities