Convolution Of Probability Distributions
   HOME

TheInfoList



OR:

The convolution/sum of probability distributions arises in
probability theory Probability theory is the branch of mathematics concerned with probability. Although there are several different probability interpretations, probability theory treats the concept in a rigorous mathematical manner by expressing it through a set o ...
and
statistics Statistics (from German language, German: ''wikt:Statistik#German, Statistik'', "description of a State (polity), state, a country") is the discipline that concerns the collection, organization, analysis, interpretation, and presentation of ...
as the operation in terms of
probability distribution In probability theory and statistics, a probability distribution is the mathematical function that gives the probabilities of occurrence of different possible outcomes for an experiment. It is a mathematical description of a random phenomenon i ...
s that corresponds to the addition of
independent Independent or Independents may refer to: Arts, entertainment, and media Artist groups * Independents (artist group), a group of modernist painters based in the New Hope, Pennsylvania, area of the United States during the early 1930s * Independ ...
random variable A random variable (also called random quantity, aleatory variable, or stochastic variable) is a mathematical formalization of a quantity or object which depends on random events. It is a mapping or a function from possible outcomes (e.g., the po ...
s and, by extension, to forming linear combinations of random variables. The operation here is a special case of
convolution In mathematics (in particular, functional analysis), convolution is a operation (mathematics), mathematical operation on two function (mathematics), functions ( and ) that produces a third function (f*g) that expresses how the shape of one is ...
in the context of probability distributions.


Introduction

The
probability distribution In probability theory and statistics, a probability distribution is the mathematical function that gives the probabilities of occurrence of different possible outcomes for an experiment. It is a mathematical description of a random phenomenon i ...
of the sum of two or more
independent Independent or Independents may refer to: Arts, entertainment, and media Artist groups * Independents (artist group), a group of modernist painters based in the New Hope, Pennsylvania, area of the United States during the early 1930s * Independ ...
random variable A random variable (also called random quantity, aleatory variable, or stochastic variable) is a mathematical formalization of a quantity or object which depends on random events. It is a mapping or a function from possible outcomes (e.g., the po ...
s is the convolution of their individual distributions. The term is motivated by the fact that the
probability mass function In probability and statistics, a probability mass function is a function that gives the probability that a discrete random variable is exactly equal to some value. Sometimes it is also known as the discrete density function. The probability mass ...
or
probability density function In probability theory, a probability density function (PDF), or density of a continuous random variable, is a function whose value at any given sample (or point) in the sample space (the set of possible values taken by the random variable) can ...
of a sum of independent random variables is the
convolution In mathematics (in particular, functional analysis), convolution is a operation (mathematics), mathematical operation on two function (mathematics), functions ( and ) that produces a third function (f*g) that expresses how the shape of one is ...
of their corresponding probability mass functions or probability density functions respectively. Many well known distributions have simple convolutions: see
List of convolutions of probability distributions In probability theory, the probability distribution of the sum of two or more independent (probability), independent random variables is the convolution of their individual distributions. The term is motivated by the fact that the probability mass ...
The general formula for the distribution of the sum Z=X+Y of two independent integer-valued (and hence discrete) random variables is Susan Holmes (1998). Sums of Random Variables: Statistics 116. Stanford. http://statweb.stanford.edu/~susan/courses/s116/node114.html :P(Z=z) = \sum_^\infty P(X=k)P(Y=z-k) For independent, continuous random variables with
probability density function In probability theory, a probability density function (PDF), or density of a continuous random variable, is a function whose value at any given sample (or point) in the sample space (the set of possible values taken by the random variable) can ...
s (PDF) f,g and
cumulative distribution function In probability theory and statistics, the cumulative distribution function (CDF) of a real-valued random variable X, or just distribution function of X, evaluated at x, is the probability that X will take a value less than or equal to x. Ev ...
s (CDF) F,G respectively, we have that the CDF of the sum is: :H(z)=\int_^\infty F(z-t)g(t) dt = \int_^\infty G(t)f(z-t) dt If we start with random variables X and Y, related by Z = X + Y, and with no information about their possible independence, then: :f_Z(z) = \int \limits_^ f_(x, z-x)~dx However, if X and Y are independent, then: :f_(x,y) = f_X(x) f_Y(y) and this formula becomes the convolution of probability distributions: :f_Z(z) = \int \limits_^ f_(x)~f_Y(z-x)~dx


Example derivation

There are several ways of deriving formulae for the convolution of probability distributions. Often the manipulation of integrals can be avoided by use of some type of
generating function In mathematics, a generating function is a way of encoding an infinite sequence of numbers () by treating them as the coefficients of a formal power series. This series is called the generating function of the sequence. Unlike an ordinary seri ...
. Such methods can also be useful in deriving properties of the resulting distribution, such as moments, even if an explicit formula for the distribution itself cannot be derived. One of the straightforward techniques is to use characteristic functions, which always exists and are unique to a given distribution.


Convolution of Bernoulli distributions

The convolution of two independent identically distributed Bernoulli random variables is a binomial random variable. That is, in a shorthand notation, : \sum_^2 \mathrm(p) \sim \mathrm(2,p) To show this let :X_i \sim \mathrm(p), \quad 0 and define :Y=\sum_^2 X_i Also, let ''Z'' denote a generic binomial random variable: :Z \sim \mathrm(2,p) \,\!


Using probability mass functions

As X_1 \text X_2 are independent, :\begin\mathbb =n=\mathbb\left sum_^2 X_i=n\right\\ &=\sum_ \mathbb _1=mtimes\mathbb _2=n-m\\ &=\sum_\left binomp^m\left(1-p\right)^\rightleft binomp^\left(1-p\right)^\right\ &=p^n\left(1-p\right)^\sum_\binom\binom \\ &=p^n\left(1-p\right)^\left binom\binom+\binom\binom\right\ &=\binomp^n\left(1-p\right)^=\mathbb =n\end Here, we used the fact that \tbinom=0 for ''k''>''n'' in the last but three equality, and of
Pascal's rule In mathematics, Pascal's rule (or Pascal's formula) is a combinatorial identity about binomial coefficients. It states that for positive natural numbers ''n'' and ''k'', + = , where \tbinom is a binomial coefficient; one interpretation of the ...
in the second last equality.


Using characteristic functions

The characteristic function of each X_k and of Z is :\varphi_(t)=1-p+pe^ \qquad \varphi_Z(t)=\left(1-p+pe^\right)^2 where ''t'' is within some
neighborhood A neighbourhood (British English, Irish English, Australian English and Canadian English) or neighborhood (American English; see spelling differences) is a geographically localised community within a larger city, town, suburb or rural area, ...
of zero. :\begin\varphi_Y(t)&=\operatorname\left(e^\right)=\operatorname\left(\prod_^2 e^\right)\\ &=\prod_^2 \operatorname\left(e^\right)=\prod_^2 \left(1-p+pe^\right)\\ &=\left(1-p+pe^\right)^2=\varphi_Z(t)\end The expectation of the product is the product of the expectations since each X_k is independent. Since Y and Z have the same characteristic function, they must have the same distribution.


See also

*
List of convolutions of probability distributions In probability theory, the probability distribution of the sum of two or more independent (probability), independent random variables is the convolution of their individual distributions. The term is motivated by the fact that the probability mass ...


References

* {{cite book , last1=Hogg , first1=Robert V. , authorlink1=Robert V. Hogg , last2=McKean , first2=Joseph W. , last3=Craig , first3=Allen T. , title=Introduction to mathematical statistics , edition=6th , publisher=Prentice Hall , url=http://www.pearsonhighered.com/educator/product/Introduction-to-Mathematical-Statistics/9780130085078.page , location=Upper Saddle River, New Jersey , year=2004 , pages=692 , ISBN=978-0-13-008507-8, MR=467974 Theory of probability distributions