De Moivre–Laplace theorem
   HOME

TheInfoList



OR:

In
probability theory Probability theory is the branch of mathematics concerned with probability. Although there are several different probability interpretations, probability theory treats the concept in a rigorous mathematical manner by expressing it through a set ...
, the de Moivre–Laplace theorem, which is a special case of the
central limit theorem In probability theory, the central limit theorem (CLT) establishes that, in many situations, when independent random variables are summed up, their properly normalized sum tends toward a normal distribution even if the original variables themsel ...
, states that the
normal distribution In statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real-valued random variable. The general form of its probability density function is : f(x) = \frac e^ The parameter \mu ...
may be used as an approximation to the
binomial distribution In probability theory and statistics, the binomial distribution with parameters ''n'' and ''p'' is the discrete probability distribution of the number of successes in a sequence of ''n'' independent experiments, each asking a yes–no ques ...
under certain conditions. In particular, the theorem shows that the
probability mass function In probability and statistics, a probability mass function is a function that gives the probability that a discrete random variable is exactly equal to some value. Sometimes it is also known as the discrete density function. The probability mass ...
of the random number of "successes" observed in a series of n
independent Independent or Independents may refer to: Arts, entertainment, and media Artist groups * Independents (artist group), a group of modernist painters based in the New Hope, Pennsylvania, area of the United States during the early 1930s * Independe ...
Bernoulli trial In the theory of probability and statistics, a Bernoulli trial (or binomial trial) is a random experiment with exactly two possible outcomes, "success" and "failure", in which the probability of success is the same every time the experiment is c ...
s, each having probability p of success (a binomial distribution with n trials), converges to the
probability density function In probability theory, a probability density function (PDF), or density of a continuous random variable, is a function whose value at any given sample (or point) in the sample space (the set of possible values taken by the random variable) ca ...
of the normal distribution with mean np and standard deviation \sqrt, as n grows large, assuming p is not 0 or 1. The theorem appeared in the second edition of ''
The Doctrine of Chances ''The Doctrine of Chances'' was the first textbook on probability theory, written by 18th-century French mathematician Abraham de Moivre and first published in 1718.. De Moivre wrote in English because he resided in England at the time, having ...
'' by
Abraham de Moivre Abraham de Moivre FRS (; 26 May 166727 November 1754) was a French mathematician known for de Moivre's formula, a formula that links complex numbers and trigonometry, and for his work on the normal distribution and probability theory. He move ...
, published in 1738. Although de Moivre did not use the term "Bernoulli trials", he wrote about the
probability distribution In probability theory and statistics, a probability distribution is the mathematical function that gives the probabilities of occurrence of different possible outcomes for an experiment. It is a mathematical description of a random phenomenon ...
of the number of times "heads" appears when a coin is tossed 3600 times. This is one derivation of the particular
Gaussian function In mathematics, a Gaussian function, often simply referred to as a Gaussian, is a function of the base form f(x) = \exp (-x^2) and with parametric extension f(x) = a \exp\left( -\frac \right) for arbitrary real constants , and non-zero . It is ...
used in the normal distribution. It is a special case of the central limit theorem because a Bernoulli process can be thought of as the drawing of independent random variables from a bimodal discrete distribution with non-zero probability only for values 0 and 1. In this case, the binomial distribution models the number of successes (i.e., the number of 1s), whereas the central limit theorem states that, given sufficiently large ''n'', the distribution of the sample means will be approximately normal. However, because in this case the fraction of successes (i.e., the number of 1s divided by the number of trials, ''n'') is ''equal to the sample mean'', the distribution of the fractions of successes (described by the binomial distribution divided by the constant ''n'') and the distribution of the sample means (approximately normal with large ''n'' due to the central limit theorem) are equivalent.


Theorem

As ''n'' grows large, for ''k'' in the
neighborhood A neighbourhood (British English, Irish English, Australian English and Canadian English) or neighborhood (American English; see spelling differences) is a geographically localised community within a larger city, town, suburb or rural area, ...
of ''np'' we can approximate : \, p^k q^ \simeq \frac\,e^, \qquad p+q=1,\ p, q > 0 in the sense that the ratio of the left-hand side to the right-hand side converges to 1 as ''n'' → ∞.


Proof

The theorem can be more rigorously stated as follows: \left(X\!\,-\!\, np\right)\!/\!\sqrt, with \textstyle X a binomially distributed random variable, approaches the standard normal as n\!\to\!\infty, with the ratio of the probability mass of X to the limiting normal density being 1. This can be shown for an arbitrary nonzero and finite point c. On the unscaled curve for X, this would be a point k given by :k=np+c\sqrt For example, with c at 3, k stays 3 standard deviations from the mean in the unscaled curve. The normal distribution with mean \mu and standard deviation \sigma is defined by the differential equation (DE) :f'\!(x)\!=\!-\!\,\fracf(x) with an initial condition set by the probability axiom \int_^\!f(x)\,dx\!=\!1. The binomial distribution limit approaches the normal if the binomial satisfies this DE. As the binomial is discrete the equation starts as a
difference equation In mathematics, a recurrence relation is an equation according to which the nth term of a sequence of numbers is equal to some combination of the previous terms. Often, only k previous terms of the sequence appear in the equation, for a parameter ...
whose limit morphs to a DE. Difference equations use th
discrete derivative
\textstyle p(k\!+\!1)\!-\!p(k), the change for step size 1. As \textstyle n\!\to\!\infty, the discrete derivative becomes the continuous derivative. Hence the proof need show only that, for the unscaled binomial distribution, :\frac\!\cdot\! \left(-\frac\right) \!\to\! 1 as n\!\to\!\infty. The required result can be shown directly: : \begin \frac\frac\!&=\frac\frac \\ &= \frac\frac \\ &= \frac\frac \\ & \to 1 \end The last holds because the term -cnpq dominates both the denominator and the numerator as n\!\to\!\infty. As \textstyle k takes just integral values, the constant \textstyle c is subject to a rounding error. However, the maximum of this error, \textstyle /\!\sqrt, is a vanishing value.


Alternate proof

The proof consists of transforming the left-hand side (in the statement of the theorem) to the right-hand side by three approximations. First, according to Stirling's formula, the factorial of a large number ''n'' can be replaced with the approximation : n! \simeq n^n e^\sqrt\qquad \text n \to \infty. Thus : \begin p^k q^ & = \frac p^k q^ \\& \simeq \frac p^k q^\\&=\sqrt\fracp^kq^\\&=\sqrt\left(\frac\right)^k\left(\frac\right)^\end Next, the approximation \tfrac \to p is used to match the root above to the desired root on the right-hand side. : \begin p^k q^ & \simeq \sqrt\left(\frac\right)^ \left(\frac\right)^\\&\simeq\frac\left(\frac\right)^ \left(\frac\right)^ \qquad p+q=1\\ \end Finally, the expression is rewritten as an exponential and the Taylor Series approximation for ln(1+x) is used: : \ln\left(1+x\right)\simeq x-\frac+\frac-\cdots Then : \begin p^k q^ &\simeq \frac\exp\left\\\ &=\frac\exp\left\\\&=\frac\exp\left\\\&=\frac\exp\left\\qquad p+q=1\\&=\frac\exp\left\\\&=\frac\exp\left\\\&=\frac\exp\left\\\&=\frac\exp\left\\\&=\frac\exp\left\\\&=\frac\exp\left\\\&\simeq\frac\exp\left\\\&=\frace^\frac\\ \end Each "\simeq" in the above argument is a statement that two quantities are asymptotically equivalent as ''n'' increases, in the same sense as in the original statement of the theorem—i.e., that the ratio of each pair of quantities approaches 1 as ''n'' → ∞.


See also

*
Poisson distribution In probability theory and statistics, the Poisson distribution is a discrete probability distribution that expresses the probability of a given number of events occurring in a fixed interval of time or space if these events occur with a known co ...
is an alternative approximation of the binomial distribution for large values of ''n''.


Notes

{{DEFAULTSORT:De Moivre-Laplace, Theorem Of Central limit theorem