HOME

TheInfoList



OR:

In mathematics, the Marcinkiewicz–Zygmund inequality, named after
Józef Marcinkiewicz Józef Marcinkiewicz (; 30 March 1910 in Cimoszka, near Białystok, Poland – 1940 in Katyn, USSR) was a Polish mathematician. He was a student of Antoni Zygmund; and later worked with Juliusz Schauder, Stefan Kaczmarz and Raphaël Salem. H ...
and
Antoni Zygmund Antoni Zygmund (December 25, 1900 – May 30, 1992) was a Polish mathematician. He worked mostly in the area of mathematical analysis, including especially harmonic analysis, and he is considered one of the greatest analysts of the 20th century. ...
, gives relations between moments of a collection of
independent random variables Independent or Independents may refer to: Arts, entertainment, and media Artist groups * Independents (artist group), a group of modernist painters based in the New Hope, Pennsylvania, area of the United States during the early 1930s * Independen ...
. It is a generalization of the rule for the sum of
variance In probability theory and statistics, variance is the expectation of the squared deviation of a random variable from its population mean or sample mean. Variance is a measure of dispersion, meaning it is a measure of how far a set of number ...
s of independent random variables to moments of arbitrary order. It is a special case of the Burkholder-Davis-Gundy inequality in the case of discrete-time martingales.


Statement of the inequality

Theorem J. Marcinkiewicz and A. Zygmund. Sur les fonctions indépendantes. ''Fund. Math.'', 28:60–90, 1937. Reprinted in Józef Marcinkiewicz, ''Collected papers'', edited by Antoni Zygmund, Panstwowe Wydawnictwo Naukowe, Warsaw, 1964, pp. 233–259. Yuan Shih Chow and Henry Teicher. ''Probability theory. Independence, interchangeability, martingales''. Springer-Verlag, New York, second edition, 1988. If \textstyle X_, \textstyle i=1,\ldots,n, are independent random variables such that \textstyle E\left( X_\right) =0 and \textstyle E\left( \left\vert X_\right\vert ^\right) <+\infty, \textstyle 1\leq p<+\infty, then : A_E\left( \left( \sum_^\left\vert X_\right\vert ^\right) _^\right) \leq E\left( \left\vert \sum_^X_\right\vert ^\right) \leq B_E\left( \left( \sum_^\left\vert X_\right\vert ^\right) _^\right) where \textstyle A_ and \textstyle B_ are positive constants, which depend only on \textstyle p and not on the underlying distribution of the random variables involved.


The second-order case

In the case \textstyle p=2, the inequality holds with \textstyle A_=B_=1, and it reduces to the rule for the sum of variances of independent random variables with zero mean, known from elementary statistics: If \textstyle E\left( X_\right) =0 and \textstyle E\left( \left\vert X_\right\vert ^\right) <+\infty, then : \mathrm\left(\sum_^X_\right)=E\left( \left\vert \sum_^X_\right\vert ^\right) =\sum_^\sum_^E\left( X_\overline_\right) =\sum_^E\left( \left\vert X_\right\vert ^\right) =\sum_^\mathrm\left(X_\right).


See also

Several similar moment inequalities are known as
Khintchine inequality In mathematics, the Khintchine inequality, named after Aleksandr Khinchin and spelled in multiple ways in the Latin alphabet, is a theorem from probability, and is also frequently used in mathematical analysis, analysis. Heuristically, it says that ...
and Rosenthal inequalities, and there are also extensions to more general symmetric
statistic A statistic (singular) or sample statistic is any quantity computed from values in a sample which is considered for a statistical purpose. Statistical purposes include estimating a population parameter, describing a sample, or evaluating a hy ...
s of independent random variables.R. Ibragimov and Sh. Sharakhmetov. Analogues of Khintchine, Marcinkiewicz–Zygmund and Rosenthal inequalities for symmetric statistics. ''Scandinavian Journal of Statistics'', 26(4):621–633, 1999.


Notes

{{DEFAULTSORT:Marcinkiewicz-Zygmund inequality Statistical inequalities Probabilistic inequalities Probability theorems Theorems in functional analysis