In
probability theory
Probability theory or probability calculus is the branch of mathematics concerned with probability. Although there are several different probability interpretations, probability theory treats the concept in a rigorous mathematical manner by expre ...
and
statistics
Statistics (from German language, German: ', "description of a State (polity), state, a country") is the discipline that concerns the collection, organization, analysis, interpretation, and presentation of data. In applying statistics to a s ...
, a normal distribution or Gaussian distribution is a type of
continuous probability distribution for a
real-valued random variable
A random variable (also called random quantity, aleatory variable, or stochastic variable) is a Mathematics, mathematical formalization of a quantity or object which depends on randomness, random events. The term 'random variable' in its mathema ...
. The general form of its
probability density function is
The parameter is the
mean or
expectation of the distribution (and also its
median
The median of a set of numbers is the value separating the higher half from the lower half of a Sample (statistics), data sample, a statistical population, population, or a probability distribution. For a data set, it may be thought of as the “ ...
and
mode), while the parameter
is the
variance
In probability theory and statistics, variance is the expected value of the squared deviation from the mean of a random variable. The standard deviation (SD) is obtained as the square root of the variance. Variance is a measure of dispersion ...
. The
standard deviation
In statistics, the standard deviation is a measure of the amount of variation of the values of a variable about its Expected value, mean. A low standard Deviation (statistics), deviation indicates that the values tend to be close to the mean ( ...
of the distribution is (sigma). A random variable with a Gaussian distribution is said to be normally distributed, and is called a normal deviate.
Normal distributions are important in
statistics
Statistics (from German language, German: ', "description of a State (polity), state, a country") is the discipline that concerns the collection, organization, analysis, interpretation, and presentation of data. In applying statistics to a s ...
and are often used in the
natural and
social science
Social science (often rendered in the plural as the social sciences) is one of the branches of science, devoted to the study of societies and the relationships among members within those societies. The term was formerly used to refer to the ...
s to represent real-valued
random variable
A random variable (also called random quantity, aleatory variable, or stochastic variable) is a Mathematics, mathematical formalization of a quantity or object which depends on randomness, random events. The term 'random variable' in its mathema ...
s whose distributions are not known. Their importance is partly due to the
central limit theorem. It states that, under some conditions, the average of many samples (observations) of a random variable with finite mean and variance is itself a random variable—whose distribution
converges to a normal distribution as the number of samples increases. Therefore, physical quantities that are expected to be the sum of many independent processes, such as
measurement errors, often have distributions that are nearly normal.
Moreover, Gaussian distributions have some unique properties that are valuable in analytic studies. For instance, any
linear combination
In mathematics, a linear combination or superposition is an Expression (mathematics), expression constructed from a Set (mathematics), set of terms by multiplying each term by a constant and adding the results (e.g. a linear combination of ''x'' a ...
of a fixed collection of independent normal deviates is a normal deviate. Many results and methods, such as
propagation of uncertainty and
least squares parameter fitting, can be derived analytically in explicit form when the relevant variables are normally distributed.
A normal distribution is sometimes informally called a bell curve.
However, many other distributions are
bell-shaped (such as the
Cauchy,
Student's ''t'', and
logistic distributions). (For other names, see ''
Naming''.)
The
univariate probability distribution is generalized for
vectors in the
multivariate normal distribution and for matrices in the
matrix normal distribution.
Definitions
Standard normal distribution
The simplest case of a normal distribution is known as the standard normal distribution or unit normal distribution. This is a special case when
and
, and it is described by this
probability density function (or density):
The variable has a mean of 0 and a variance and standard deviation of 1. The density
has its peak
at
and
inflection points at
and .
Although the density above is most commonly known as the ''standard normal,'' a few authors have used that term to describe other versions of the normal distribution.
Carl Friedrich Gauss, for example, once defined the standard normal as
which has a variance of , and
Stephen Stigler once defined the standard normal as
which has a simple functional form and a variance of
General normal distribution
Every normal distribution is a version of the standard normal distribution, whose domain has been stretched by a factor (the standard deviation) and then translated by (the mean value):
The probability density must be scaled by
so that the
integral
In mathematics, an integral is the continuous analog of a Summation, sum, which is used to calculate area, areas, volume, volumes, and their generalizations. Integration, the process of computing an integral, is one of the two fundamental oper ...
is still 1.
If is a
standard normal deviate, then
will have a normal distribution with expected value and standard deviation . This is equivalent to saying that the standard normal distribution can be scaled/stretched by a factor of and shifted by to yield a different normal distribution, called . Conversely, if is a normal deviate with parameters and
, then this distribution can be re-scaled and shifted via the formula
to convert it to the standard normal distribution. This variate is also called the standardized form of .
Notation
The probability density of the standard Gaussian distribution (standard normal distribution, with zero mean and unit variance) is often denoted with the Greek letter (
phi). The alternative form of the Greek letter phi, , is also used quite often.
The normal distribution is often referred to as
or . Thus when a random variable is normally distributed with mean and standard deviation , one may write
Alternative parameterizations
Some authors advocate using the
precision as the parameter defining the width of the distribution, instead of the standard deviation or the variance . The precision is normally defined as the reciprocal of the variance, . The formula for the distribution then becomes
This choice is claimed to have advantages in numerical computations when is very close to zero, and simplifies formulas in some contexts, such as in the
Bayesian inference of variables with
multivariate normal distribution.
Alternatively, the reciprocal of the standard deviation
might be defined as the ''precision'', in which case the expression of the normal distribution becomes
According to Stigler, this formulation is advantageous because of a much simpler and easier-to-remember formula, and simple approximate formulas for the
quantile
In statistics and probability, quantiles are cut points dividing the range of a probability distribution into continuous intervals with equal probabilities or dividing the observations in a sample in the same way. There is one fewer quantile t ...
s of the distribution.
Normal distributions form an
exponential family with
natural parameters
and
, and natural statistics ''x'' and ''x''
2. The dual expectation parameters for normal distribution are and .
Cumulative distribution function
The
cumulative distribution function (CDF) of the standard normal distribution, usually denoted with the capital Greek letter , is the integral
Error function
The related
error function gives the probability of a random variable, with normal distribution of mean 0 and variance 1/2 falling in the range . That is:
These integrals cannot be expressed in terms of elementary functions, and are often said to be
special functions. However, many numerical approximations are known; see
below for more.
The two functions are closely related, namely
For a generic normal distribution with density , mean and variance
, the cumulative distribution function is
The complement of the standard normal cumulative distribution function,
, is often called the
Q-function, especially in engineering texts. It gives the probability that the value of a standard normal random variable will exceed : . Other definitions of the -function, all of which are simple transformations of , are also used occasionally.
The
graph of the standard normal cumulative distribution function has 2-fold
rotational symmetry around the point (0,1/2); that is, . Its
antiderivative (indefinite integral) can be expressed as follows:
The cumulative distribution function of the standard normal distribution can be expanded by
integration by parts
In calculus, and more generally in mathematical analysis, integration by parts or partial integration is a process that finds the integral of a product of functions in terms of the integral of the product of their derivative and antiderivati ...
into a series:
where
denotes the
double factorial
In mathematics, the double factorial of a number , denoted by , is the product of all the positive integers up to that have the same Parity (mathematics), parity (odd or even) as . That is,
n!! = \prod_^ (n-2k) = n (n-2) (n-4) \cdots.
Restated ...
.
An
asymptotic expansion
In mathematics, an asymptotic expansion, asymptotic series or Poincaré expansion (after Henri Poincaré) is a formal series of functions which has the property that truncating the series after a finite number of terms provides an approximation ...
of the cumulative distribution function for large ''x'' can also be derived using integration by parts. For more, see .
A quick approximation to the standard normal distribution's cumulative distribution function can be found by using a Taylor series approximation:
Recursive computation with Taylor series expansion
The recursive nature of the
family of derivatives may be used to easily construct a rapidly converging
Taylor series expansion using recursive entries about any point of known value of the distribution,
:
where:
Using the Taylor series and Newton's method for the inverse function
An application for the above Taylor series expansion is to use
Newton's method to reverse the computation. That is, if we have a value for the
cumulative distribution function,
, but do not know the x needed to obtain the
, we can use Newton's method to find x, and use the Taylor series expansion above to minimize the number of computations. Newton's method is ideal to solve this problem because the first derivative of
, which is an integral of the normal standard distribution, is the normal standard distribution, and is readily available to use in the Newton's method solution.
To solve, select a known approximate solution,
, to the desired .
may be a value from a distribution table, or an intelligent estimate followed by a computation of
using any desired means to compute. Use this value of
and the Taylor series expansion above to minimize computations.
Repeat the following process until the difference between the computed
and the desired , which we will call
, is below a chosen acceptably small error, such as 10
−5, 10
−15, etc.:
where
:
is the
from a Taylor series solution using
and
When the repeated computations converge to an error below the chosen acceptably small value, ''x'' will be the value needed to obtain a
of the desired value, .
Standard deviation and coverage

About 68% of values drawn from a normal distribution are within one standard deviation ''σ'' from the mean; about 95% of the values lie within two standard deviations; and about 99.7% are within three standard deviations.
This fact is known as the
68–95–99.7 (empirical) rule, or the ''3-sigma rule''.
More precisely, the probability that a normal deviate lies in the range between
and
is given by
To 12 significant digits, the values for
are:
For large , one can use the approximation
.
Quantile function
The
quantile function of a distribution is the inverse of the cumulative distribution function. The quantile function of the standard normal distribution is called the
probit function, and can be expressed in terms of the inverse
error function:
For a normal random variable with mean and variance
, the quantile function is
The
quantile
In statistics and probability, quantiles are cut points dividing the range of a probability distribution into continuous intervals with equal probabilities or dividing the observations in a sample in the same way. There is one fewer quantile t ...
of the standard normal distribution is commonly denoted as . These values are used in
hypothesis testing, construction of
confidence intervals and
Q–Q plots. A normal random variable will exceed
with probability
, and will lie outside the interval
with probability . In particular, the quantile
is
1.96; therefore a normal random variable will lie outside the interval
in only 5% of cases.
The following table gives the quantile
such that will lie in the range
with a specified probability . These values are useful to determine
tolerance interval for
sample averages and other statistical
estimators with normal (or
asymptotic
In analytic geometry, an asymptote () of a curve is a line such that the distance between the curve and the line approaches zero as one or both of the ''x'' or ''y'' coordinates Limit of a function#Limits at infinity, tends to infinity. In pro ...
ally normal) distributions. The following table shows
, not
as defined above.
For small , the quantile function has the useful
asymptotic expansion
In mathematics, an asymptotic expansion, asymptotic series or Poincaré expansion (after Henri Poincaré) is a formal series of functions which has the property that truncating the series after a finite number of terms provides an approximation ...
Properties
The normal distribution is the only distribution whose
cumulants beyond the first two (i.e., other than the mean and
variance
In probability theory and statistics, variance is the expected value of the squared deviation from the mean of a random variable. The standard deviation (SD) is obtained as the square root of the variance. Variance is a measure of dispersion ...
) are zero. It is also the continuous distribution with the
maximum entropy for a specified mean and variance. Geary has shown, assuming that the mean and variance are finite, that the normal distribution is the only distribution where the mean and variance calculated from a set of independent draws are independent of each other.
[Geary RC(1936) The distribution of the "Student's ratio for the non-normal samples". Supplement to the Journal of the Royal Statistical Society 3 (2): 178–184]
The normal distribution is a subclass of the
elliptical distributions. The normal distribution is
symmetric about its mean, and is non-zero over the entire real line. As such it may not be a suitable model for variables that are inherently positive or strongly skewed, such as the
weight
In science and engineering, the weight of an object is a quantity associated with the gravitational force exerted on the object by other objects in its environment, although there is some variation and debate as to the exact definition.
Some sta ...
of a person or the price of a
share. Such variables may be better described by other distributions, such as the
log-normal distribution or the
Pareto distribution
The Pareto distribution, named after the Italian civil engineer, economist, and sociologist Vilfredo Pareto, is a power-law probability distribution that is used in description of social, quality control, scientific, geophysical, actuarial scien ...
.
The value of the normal density is practically zero when the value lies more than a few
standard deviation
In statistics, the standard deviation is a measure of the amount of variation of the values of a variable about its Expected value, mean. A low standard Deviation (statistics), deviation indicates that the values tend to be close to the mean ( ...
s away from the mean (e.g., a spread of three standard deviations covers all but 0.27% of the total distribution). Therefore, it may not be an appropriate model when one expects a significant fraction of
outlier
In statistics, an outlier is a data point that differs significantly from other observations. An outlier may be due to a variability in the measurement, an indication of novel data, or it may be the result of experimental error; the latter are ...
s—values that lie many standard deviations away from the mean—and least squares and other
statistical inference methods that are optimal for normally distributed variables often become highly unreliable when applied to such data. In those cases, a more
heavy-tailed distribution should be assumed and the appropriate
robust statistical inference methods applied.
The Gaussian distribution belongs to the family of
stable distributions which are the attractors of sums of
independent, identically distributed distributions whether or not the mean or variance is finite. Except for the Gaussian which is a limiting case, all stable distributions have heavy tails and infinite variance. It is one of the few distributions that are stable and that have probability density functions that can be expressed analytically, the others being the
Cauchy distribution and the
Lévy distribution.
Symmetries and derivatives
The normal distribution with density
(mean and variance
) has the following properties:
* It is symmetric around the point
which is at the same time the
mode, the
median
The median of a set of numbers is the value separating the higher half from the lower half of a Sample (statistics), data sample, a statistical population, population, or a probability distribution. For a data set, it may be thought of as the “ ...
and the
mean of the distribution.
* It is
unimodal: its first
derivative
In mathematics, the derivative is a fundamental tool that quantifies the sensitivity to change of a function's output with respect to its input. The derivative of a function of a single variable at a chosen input value, when it exists, is t ...
is positive for
negative for
and zero only at
* The area bounded by the curve and the -axis is unity (i.e. equal to one).
* Its first derivative is
* Its second derivative is
* Its density has two
inflection points (where the second derivative of is zero and changes sign), located one standard deviation away from the mean, namely at
and
* Its density is
log-concave.
* Its density is infinitely
differentiable, indeed
supersmooth of order 2.
Furthermore, the density of the standard normal distribution (i.e.
and
) also has the following properties:
* Its first derivative is
* Its second derivative is
* More generally, its th derivative is
where
is the th (probabilist)
Hermite polynomial.
* The probability that a normally distributed variable with known and
is in a particular set, can be calculated by using the fact that the fraction
has a standard normal distribution.
Moments
The plain and absolute
moments of a variable are the expected values of
and
, respectively. If the expected value of is zero, these parameters are called ''central moments;'' otherwise, these parameters are called ''non-central moments.'' Usually we are interested only in moments with integer order .
If has a normal distribution, the non-central moments exist and are finite for any whose real part is greater than −1. For any non-negative integer , the plain central moments are:
Here
denotes the
double factorial
In mathematics, the double factorial of a number , denoted by , is the product of all the positive integers up to that have the same Parity (mathematics), parity (odd or even) as . That is,
n!! = \prod_^ (n-2k) = n (n-2) (n-4) \cdots.
Restated ...
, that is, the product of all numbers from to 1 that have the same parity as
The central absolute moments coincide with plain moments for all even orders, but are nonzero for odd orders. For any non-negative integer