In
probability theory
Probability theory is the branch of mathematics concerned with probability. Although there are several different probability interpretations, probability theory treats the concept in a rigorous mathematical manner by expressing it through a set ...
, calculation of the sum of normally distributed random variables is an instance of the arithmetic of
random variable
A random variable (also called random quantity, aleatory variable, or stochastic variable) is a mathematical formalization of a quantity or object which depends on random events. It is a mapping or a function from possible outcomes (e.g., the po ...
s, which can be quite complex based on the
probability distribution
In probability theory and statistics, a probability distribution is the mathematical function that gives the probabilities of occurrence of different possible outcomes for an experiment. It is a mathematical description of a random phenomenon ...
s of the random variables involved and their relationships.
This is not to be confused with the sum of normal distributions which forms a
mixture distribution
In probability and statistics, a mixture distribution is the probability distribution of a random variable that is derived from a collection of other random variables as follows: first, a random variable is selected by chance from the collectio ...
.
Independent random variables
Let ''X'' and ''Y'' be
independent
Independent or Independents may refer to:
Arts, entertainment, and media Artist groups
* Independents (artist group), a group of modernist painters based in the New Hope, Pennsylvania, area of the United States during the early 1930s
* Independe ...
random variable
A random variable (also called random quantity, aleatory variable, or stochastic variable) is a mathematical formalization of a quantity or object which depends on random events. It is a mapping or a function from possible outcomes (e.g., the po ...
s that are
normally distributed (and therefore also jointly so), then their sum is also normally distributed. i.e., if
:
:
:
then
:
This means that the sum of two independent normally distributed random variables is normal, with its mean being the sum of the two means, and its variance being the sum of the two variances (i.e., the square of the standard deviation is the sum of the squares of the standard deviations).
In order for this result to hold, the assumption that ''X'' and ''Y'' are independent cannot be dropped, although it can be weakened to the assumption that ''X'' and ''Y'' are
jointly, rather than separately, normally distributed. (See
here for an example.)
The result about the mean holds in all cases, while the result for the variance requires uncorrelatedness, but not independence.
Proofs
Proof using characteristic functions
The
characteristic function In mathematics, the term "characteristic function" can refer to any of several distinct concepts:
* The indicator function of a subset, that is the function
::\mathbf_A\colon X \to \,
:which for a given subset ''A'' of ''X'', has value 1 at points ...
:
of the sum of two independent random variables ''X'' and ''Y'' is just the product of the two separate characteristic functions:
:
of ''X'' and ''Y''.
The characteristic function of the normal distribution with expected value μ and variance σ
2 is
:
So
:
This is the characteristic function of the normal distribution with expected value
and variance
Finally, recall that no two distinct distributions can both have the same characteristic function, so the distribution of ''X'' + ''Y'' must be just this normal distribution.
Proof using convolutions
For independent random variables ''X'' and ''Y'', the distribution ''f''
''Z'' of ''Z'' = ''X'' + ''Y'' equals the convolution of ''f''
''X'' and ''f''
''Y'':
:
Given that ''f''
''X'' and ''f''
''Y'' are normal densities,
:
Substituting into the convolution:
:
Defining
, and
completing the square
:
In elementary algebra, completing the square is a technique for converting a quadratic polynomial of the form
:ax^2 + bx + c
to the form
:a(x-h)^2 + k
for some values of ''h'' and ''k''.
In other words, completing the square places a perfe ...
:
:
The expression in the integral is a normal density distribution on ''x'', and so the integral evaluates to 1. The desired result follows:
: