HOME

TheInfoList



OR:

In mathematics, the second moment method is a technique used in
probability theory Probability theory is the branch of mathematics concerned with probability. Although there are several different probability interpretations, probability theory treats the concept in a rigorous mathematical manner by expressing it through a set o ...
and
analysis Analysis ( : analyses) is the process of breaking a complex topic or substance into smaller parts in order to gain a better understanding of it. The technique has been applied in the study of mathematics and logic since before Aristotle (38 ...
to show that a
random variable A random variable (also called random quantity, aleatory variable, or stochastic variable) is a mathematical formalization of a quantity or object which depends on random events. It is a mapping or a function from possible outcomes (e.g., the po ...
has positive probability of being positive. More generally, the "moment method" consists of bounding the probability that a random variable fluctuates far from its mean, by using its moments. The method is often quantitative, in that one can often deduce a lower bound on the probability that the random variable is larger than some constant times its expectation. The method involves comparing the second
moment Moment or Moments may refer to: * Present time Music * The Moments, American R&B vocal group Albums * ''Moment'' (Dark Tranquillity album), 2020 * ''Moment'' (Speed album), 1998 * ''Moments'' (Darude album) * ''Moments'' (Christine Guldbrand ...
of random variables to the square of the first moment.


First moment method

The first moment method is a simple application of
Markov's inequality In probability theory, Markov's inequality gives an upper bound for the probability that a non-negative function (mathematics), function of a random variable is greater than or equal to some positive Constant (mathematics), constant. It is named a ...
for integer-valued variables. For a non-negative, integer-valued random variable ''X'', we may want to prove that ''X'' = 0 with high probability. To obtain an upper bound for P(''X'' > 0), and thus a lower bound for P(''X'' = 0), we first note that since ''X'' takes only integer values, P(''X'' > 0) = P(''X'' ≥ 1). Since ''X'' is non-negative we can now apply
Markov's inequality In probability theory, Markov's inequality gives an upper bound for the probability that a non-negative function (mathematics), function of a random variable is greater than or equal to some positive Constant (mathematics), constant. It is named a ...
to obtain P(''X'' ≥ 1) ≤ E 'X'' Combining these we have P(''X'' > 0) ≤ E 'X'' the first moment method is simply the use of this inequality.


Second moment method

In the other direction, E 'X''being "large" does not directly imply that P(''X'' = 0) is small. However, we can often use the second moment to derive such a conclusion, using
Cauchy–Schwarz inequality The Cauchy–Schwarz inequality (also called Cauchy–Bunyakovsky–Schwarz inequality) is considered one of the most important and widely used inequalities in mathematics. The inequality for sums was published by . The corresponding inequality fo ...
. Theorem: If ''X'' ≥ 0 is a
random variable A random variable (also called random quantity, aleatory variable, or stochastic variable) is a mathematical formalization of a quantity or object which depends on random events. It is a mapping or a function from possible outcomes (e.g., the po ...
with finite variance, then : \operatorname( X > 0 ) \ge \frac. Proof: Using the
Cauchy–Schwarz inequality The Cauchy–Schwarz inequality (also called Cauchy–Bunyakovsky–Schwarz inequality) is considered one of the most important and widely used inequalities in mathematics. The inequality for sums was published by . The corresponding inequality fo ...
, we have : \operatorname = \operatorname X \, \mathbf_ \le \operatorname ^2 \operatorname( X > 0)^. Solving for \operatorname( X > 0), the desired inequality then follows. ∎ The method can also be used on distributional limits of random variables. Furthermore, the estimate of the previous theorem can be refined by means of the so-called
Paley–Zygmund inequality In mathematics, the Paley–Zygmund inequality bounds the probability that a positive random variable is small, in terms of its first two moments. The inequality was proved by Raymond Paley and Antoni Zygmund. Theorem: If ''Z'' ≥ 0 is ...
. Suppose that ''Xn'' is a sequence of non-negative real-valued random variables which converge in law to a random variable ''X''. If there are finite positive constants ''c''1, ''c''2 such that :\operatorname \left _n^2 \right \le c_1 \operatorname _n2 :\operatorname \left _n \right ge c_2 hold for every ''n'', then it follows from the
Paley–Zygmund inequality In mathematics, the Paley–Zygmund inequality bounds the probability that a positive random variable is small, in terms of its first two moments. The inequality was proved by Raymond Paley and Antoni Zygmund. Theorem: If ''Z'' ≥ 0 is ...
that for every ''n'' and θ in (0, 1) : \operatorname (X_n \geq c_2 \theta) \geq \frac. Consequently, the same inequality is satisfied by ''X''.


Example application of method


Setup of problem

The Bernoulli bond percolation subgraph of a graph ''G'' at parameter ''p'' is a random subgraph obtained from ''G'' by deleting every edge of ''G'' with probability 1−''p'', independently. The infinite complete binary tree ''T'' is an infinite
tree In botany, a tree is a perennial plant with an elongated stem, or trunk, usually supporting branches and leaves. In some usages, the definition of a tree may be narrower, including only woody plants with secondary growth, plants that are ...
where one vertex (called the root) has two neighbors and every other vertex has three neighbors. The second moment method can be used to show that at every parameter ''p'' ∈ (1/2, 1] with positive probability the connected component of the root in the percolation subgraph of ''T'' is infinite.


Application of method

Let ''K'' be the percolation component of the root, and let ''Tn'' be the set of vertices of ''T'' that are at distance ''n'' from the root. Let ''Xn'' be the number of vertices in ''Tn'' ∩ ''K''. To prove that ''K'' is infinite with positive probability, it is enough to show that \limsup_1_>0 with positive probability. By the
reverse Fatou lemma In mathematics, Fatou's lemma establishes an inequality relating the Lebesgue integral of the limit inferior of a sequence of functions to the limit inferior of integrals of these functions. The lemma is named after Pierre Fatou. Fatou's lemma ...
, it suffices to show that \inf_P(X_n>0)>0. The
Cauchy–Schwarz inequality The Cauchy–Schwarz inequality (also called Cauchy–Bunyakovsky–Schwarz inequality) is considered one of the most important and widely used inequalities in mathematics. The inequality for sums was published by . The corresponding inequality fo ...
gives :E _n2\le E _n^2\, E\left 1_)^2\right E _n^2,P(X_n>0). Therefore, it is sufficient to show that : \inf_n \frac>0\,, that is, that the second moment is bounded from above by a constant times the first moment squared (and both are nonzero). In many applications of the second moment method, one is not able to calculate the moments precisely, but can nevertheless establish this inequality. In this particular application, these moments can be calculated. For every specific ''v'' in ''Tn'', :P(v\in K)=p^n. Since , T_n, =2^n, it follows that :E _n2^n\,p^n which is the first moment. Now comes the second moment calculation. :E\!\left _n^2 \right = E\!\left sum_ \sum_1_\,1_\right= \sum_ \sum_ P(v,u\in K). For each pair ''v'', ''u'' in ''Tn'' let ''w(v, u)'' denote the vertex in ''T'' that is farthest away from the root and lies on the simple path in ''T'' to each of the two vertices ''v'' and ''u'', and let ''k(v, u)'' denote the distance from ''w'' to the root. In order for ''v'', ''u'' to both be in ''K'', it is necessary and sufficient for the three simple paths from ''w(v, u)'' to ''v'', ''u'' and the root to be in ''K''. Since the number of edges contained in the union of these three paths is 2''n'' − ''k(v, u)'', we obtain :P(v,u\in K)= p^. The number of pairs ''(v, u)'' such that ''k(v, u)'' = ''s'' is equal to 2^s\,2^\,2^=2^, for ''s'' = 0, 1, ..., ''n''. Hence, :E\!\left _n^2 \right = \sum_^n 2^ p^ = \frac 12\,(2p)^n \sum_^n (2p)^s = \frac12\,(2p)^n \, \frac \le \frac p \,E _n2, which completes the proof.


Discussion

*The choice of the random variables ''X''''n'' was rather natural in this setup. In some more difficult applications of the method, some ingenuity might be required in order to choose the random variables ''X''''n'' for which the argument can be carried through. *The
Paley–Zygmund inequality In mathematics, the Paley–Zygmund inequality bounds the probability that a positive random variable is small, in terms of its first two moments. The inequality was proved by Raymond Paley and Antoni Zygmund. Theorem: If ''Z'' ≥ 0 is ...
is sometimes used instead of the
Cauchy–Schwarz inequality The Cauchy–Schwarz inequality (also called Cauchy–Bunyakovsky–Schwarz inequality) is considered one of the most important and widely used inequalities in mathematics. The inequality for sums was published by . The corresponding inequality fo ...
and may occasionally give more refined results. *Under the (incorrect) assumption that the events ''v'', ''u'' in ''K'' are always independent, one has P(v,u\in K)=P(v\in K)\,P(u\in K), and the second moment is equal to the first moment squared. The second moment method typically works in situations in which the corresponding events or random variables are “nearly independent". *In this application, the random variables ''X''''n'' are given as sums :: X_n=\sum_1_. : In other applications, the corresponding useful random variables are
integral In mathematics Mathematics is an area of knowledge that includes the topics of numbers, formulas and related structures, shapes and the spaces in which they are contained, and quantities and their changes. These topics are represented i ...
s :: X_n=\int f_n(t)\,d\mu(t), : where the functions ''f''''n'' are random. In such a situation, one considers the product measure ''μ'' × ''μ'' and calculates :: \begin E \left _n^2 \right & = E\left int\int f_n(x)\,f_n(y)\,d\mu(x)\,d\mu(y)\right \\ & = E\left \int\int_E\left[f_n(x)\,f_n(y)\right,d\mu(x)\,d\mu(y)\right_.html" ;"title="_n(x)\,f_n(y)\right.html" ;"title="\int\int E\left[f_n(x)\,f_n(y)\right">\int\int E\left[f_n(x)\,f_n(y)\right,d\mu(x)\,d\mu(y)\right ">_n(x)\,f_n(y)\right.html" ;"title="\int\int E\left[f_n(x)\,f_n(y)\right">\int\int E\left[f_n(x)\,f_n(y)\right,d\mu(x)\,d\mu(y)\right \end :where the last step is typically justified using Fubini's theorem.


References

* * *{{Citation , last1=Lyons , first1=Russell , last2=Peres , first2=Yuval , title=Probability on trees and networks , url=http://mypage.iu.edu/~rdlyons/prbtree/prbtree.html Probabilistic inequalities Articles containing proofs Moment (mathematics)