In
probability theory
Probability theory is the branch of mathematics concerned with probability. Although there are several different probability interpretations, probability theory treats the concept in a rigorous mathematical manner by expressing it through a set ...
and
statistics, the binomial distribution with parameters ''n'' and ''p'' is the
discrete probability distribution
In probability theory and statistics, a probability distribution is the mathematical function that gives the probabilities of occurrence of different possible outcomes for an experiment. It is a mathematical description of a random phenomenon ...
of the number of successes in a sequence of ''n''
independent
Independent or Independents may refer to:
Arts, entertainment, and media Artist groups
* Independents (artist group), a group of modernist painters based in the New Hope, Pennsylvania, area of the United States during the early 1930s
* Independ ...
experiment
An experiment is a procedure carried out to support or refute a hypothesis, or determine the efficacy or likelihood of something previously untried. Experiments provide insight into Causality, cause-and-effect by demonstrating what outcome oc ...
s, each asking a
yes–no question
In linguistics, a yes–no question, also known as a binary question, a polar question, or a general question is a question whose expected answer is one of two choices, one that provides an affirmative answer to the question versus one that provid ...
, and each with its own
Boolean-valued
outcome: ''success'' (with probability ''p'') or ''failure'' (with probability
). A single success/failure experiment is also called a
Bernoulli trial
In the theory of probability and statistics, a Bernoulli trial (or binomial trial) is a random experiment with exactly two possible outcomes, "success" and "failure", in which the probability of success is the same every time the experiment is ...
or Bernoulli experiment, and a sequence of outcomes is called a
Bernoulli process; for a single trial, i.e., ''n'' = 1, the binomial distribution is a
Bernoulli distribution
In probability theory and statistics, the Bernoulli distribution, named after Swiss mathematician Jacob Bernoulli,James Victor Uspensky: ''Introduction to Mathematical Probability'', McGraw-Hill, New York 1937, page 45 is the discrete probabi ...
. The binomial distribution is the basis for the popular
binomial test
In statistics, the binomial test is an exact test of the statistical significance of deviations from a theoretically expected distribution of observations into two categories using sample data.
Usage
The binomial test is useful to test hypoth ...
of
statistical significance.
The binomial distribution is frequently used to model the number of successes in a sample of size ''n'' drawn
with replacement from a population of size ''N''. If the sampling is carried out without replacement, the draws are not independent and so the resulting distribution is a
hypergeometric distribution
In probability theory and statistics, the hypergeometric distribution is a discrete probability distribution that describes the probability of k successes (random draws for which the object drawn has a specified feature) in n draws, ''without'' ...
, not a binomial one. However, for ''N'' much larger than ''n'', the binomial distribution remains a good approximation, and is widely used.
Definitions
Probability mass function
In general, if the
random variable ''X'' follows the binomial distribution with parameters ''n''
∈
In mathematics, an element (or member) of a Set (mathematics), set is any one of the Equality (mathematics), distinct Mathematical object, objects that belong to that set.
Sets
Writing A = \ means that the elements of the set are the numbers 1, ...
and ''p'' ∈
,1 we write ''X'' ~ B(''n'', ''p''). The probability of getting exactly ''k'' successes in ''n'' independent Bernoulli trials is given by the
probability mass function:
:
for ''k'' = 0, 1, 2, ..., ''n'', where
:
is the
binomial coefficient, hence the name of the distribution. The formula can be understood as follows: ''k'' successes occur with probability ''p''
''k'' and ''n'' − ''k'' failures occur with probability
. However, the ''k'' successes can occur anywhere among the ''n'' trials, and there are
different ways of distributing ''k'' successes in a sequence of ''n'' trials.
In creating reference tables for binomial distribution probability, usually the table is filled in up to ''n''/2 values. This is because for ''k'' > ''n''/2, the probability can be calculated by its complement as
:
Looking at the expression ''f''(''k'', ''n'', ''p'') as a function of ''k'', there is a ''k'' value that maximizes it. This ''k'' value can be found by calculating
:
and comparing it to 1. There is always an integer ''M'' that satisfies
:
''f''(''k'', ''n'', ''p'') is monotone increasing for ''k'' < ''M'' and monotone decreasing for ''k'' > ''M'', with the exception of the case where (''n'' + 1)''p'' is an integer. In this case, there are two values for which ''f'' is maximal: (''n'' + 1)''p'' and (''n'' + 1)''p'' − 1. ''M'' is the ''most probable'' outcome (that is, the most likely, although this can still be unlikely overall) of the Bernoulli trials and is called the
mode
Mode ( la, modus meaning "manner, tune, measure, due measure, rhythm, melody") may refer to:
Arts and entertainment
* '' MO''D''E (magazine)'', a defunct U.S. women's fashion magazine
* ''Mode'' magazine, a fictional fashion magazine which is ...
.
Example
Suppose a
biased coin
In probability theory and statistics, a sequence of independent Bernoulli trials with probability 1/2 of success on each trial is metaphorically called a fair coin. One for which the probability is not 1/2 is called a biased or unfair coin. In the ...
comes up heads with probability 0.3 when tossed. The probability of seeing exactly 4 heads in 6 tosses is
:
Cumulative distribution function
The
cumulative distribution function can be expressed as:
:
where
is the "floor" under ''k'', i.e. the
greatest integer less than or equal to ''k''.
It can also be represented in terms of the
regularized incomplete beta function, as follows:
:
which is equivalent to the
cumulative distribution function of the
-distribution:
:
Some closed-form bounds for the cumulative distribution function are given
below.
Properties
Expected value and variance
If ''X'' ~ ''B''(''n'', ''p''), that is, ''X'' is a binomially distributed random variable, ''n'' being the total number of experiments and ''p'' the probability of each experiment yielding a successful result, then the
expected value of ''X'' is:
:
This follows from the linearity of the expected value along with the fact that is the sum of identical Bernoulli random variables, each with expected value . In other words, if
are identical (and independent) Bernoulli random variables with parameter , then
and
:
The
variance
In probability theory and statistics, variance is the expectation of the squared deviation of a random variable from its population mean or sample mean. Variance is a measure of dispersion, meaning it is a measure of how far a set of numbe ...
is:
:
This similarly follows from the fact that the variance of a sum of independent random variables is the sum of the variances.
Higher moments
The first 6
central moment
In probability theory and statistics, a central moment is a moment of a probability distribution of a random variable about the random variable's mean; that is, it is the expected value of a specified integer power of the deviation of the random ...
s, defined as
, are given by
:
The non-central moments satisfy
:
and in general
:
where
are the
Stirling numbers of the second kind
In mathematics, particularly in combinatorics, a Stirling number of the second kind (or Stirling partition number) is the number of ways to partition a set of ''n'' objects into ''k'' non-empty subsets and is denoted by S(n,k) or \textstyle \lef ...
, and
is the
th
falling power of
.
A simple bound
follows by bounding the Binomial moments via the
higher Poisson moments:
::
This shows that if
, then