In
information theory, the entropy power inequality (EPI) is a result that relates to so-called "entropy power" of
random variables. It shows that the entropy power of suitably
well-behaved
In mathematics, when a mathematical phenomenon runs counter to some intuition, then the phenomenon is sometimes called pathological. On the other hand, if a phenomenon does not run counter to intuition,
it is sometimes called well-behaved. Th ...
random variables is a
superadditive function
Function or functionality may refer to:
Computing
* Function key, a type of key on computer keyboards
* Function model, a structured representation of processes in a system
* Function object or functor or functionoid, a concept of object-oriente ...
. The entropy power inequality was proved in 1948 by
Claude Shannon
Claude Elwood Shannon (April 30, 1916 – February 24, 2001) was an American mathematician, electrical engineer, and cryptographer known as a "father of information theory".
As a 21-year-old master's degree student at the Massachusetts Inst ...
in his seminal paper "
A Mathematical Theory of Communication
"A Mathematical Theory of Communication" is an article by mathematician Claude E. Shannon published in ''Bell System Technical Journal'' in 1948. It was renamed ''The Mathematical Theory of Communication'' in the 1949 book of the same name, a sma ...
". Shannon also provided a sufficient condition for equality to hold; Stam (1959) showed that the condition is in fact necessary.
Statement of the inequality
For a random vector ''X'' : Ω → R
''n'' with
probability density function
In probability theory, a probability density function (PDF), or density of a continuous random variable, is a function whose value at any given sample (or point) in the sample space (the set of possible values taken by the random variable) ca ...
''f'' : R
''n'' → R, the
differential entropy
Differential entropy (also referred to as continuous entropy) is a concept in information theory that began as an attempt by Claude Shannon to extend the idea of (Shannon) entropy, a measure of average surprisal of a random variable, to continuo ...
of ''X'', denoted ''h''(''X''), is defined to be
:
and the entropy power of ''X'', denoted ''N''(''X''), is defined to be
:
In particular, ''N''(''X'') = , ''K'',
1/''n'' when ''X'' is normal distributed with covariance matrix ''K''.
Let ''X'' and ''Y'' be
independent random variables
Independent or Independents may refer to:
Arts, entertainment, and media Artist groups
* Independents (artist group), a group of modernist painters based in the New Hope, Pennsylvania, area of the United States during the early 1930s
* Independe ...
with probability density functions in the
''L''''p'' space ''L''
''p''(R
''n'') for some ''p'' > 1. Then
:
Moreover, equality holds
if and only if
In logic and related fields such as mathematics and philosophy, "if and only if" (shortened as "iff") is a biconditional logical connective between statements, where either both statements are true or both are false.
The connective is b ...
''X'' and ''Y'' are
multivariate normal
In probability theory and statistics, the multivariate normal distribution, multivariate Gaussian distribution, or joint normal distribution is a generalization of the one-dimensional (univariate) normal distribution to higher dimensions. One d ...
random variables with proportional
covariance matrices
In probability theory and statistics, a covariance matrix (also known as auto-covariance matrix, dispersion matrix, variance matrix, or variance–covariance matrix) is a square matrix giving the covariance between each pair of elements of ...
.
Alternative form of the inequality
The entropy power inequality can be rewritten in an equivalent form that does not explicitly depend on the definition of entropy power (see Costa and Cover reference below).
Let ''X'' and ''Y'' be
independent random variables
Independent or Independents may refer to:
Arts, entertainment, and media Artist groups
* Independents (artist group), a group of modernist painters based in the New Hope, Pennsylvania, area of the United States during the early 1930s
* Independe ...
, as above. Then, let X' and Y' be independently distributed random variables with gaussian distributions, such that
:
and
Then,
:
See also
*
Information entropy
In information theory, the entropy of a random variable is the average level of "information", "surprise", or "uncertainty" inherent to the variable's possible outcomes. Given a discrete random variable X, which takes values in the alphabet \ ...
*
Information theory
*
Limiting density of discrete points
In information theory, the limiting density of discrete points is an adjustment to the formula of Claude Shannon for differential entropy.
It was formulated by Edwin Thompson Jaynes to address defects in the initial definition of differential e ...
*
Self-information
In information theory, the information content, self-information, surprisal, or Shannon information is a basic quantity derived from the probability of a particular event occurring from a random variable. It can be thought of as an alternative ...
*
Kullback–Leibler divergence
In mathematical statistics, the Kullback–Leibler divergence (also called relative entropy and I-divergence), denoted D_\text(P \parallel Q), is a type of statistical distance: a measure of how one probability distribution ''P'' is different fr ...
*
Entropy estimation
References
*
*
*
*
* {{cite journal
, last = Stam
, first = A. J.
, title = Some inequalities satisfied by the quantities of information of Fisher and Shannon
, journal = Information and Control
, volume = 2
, year = 1959
, pages = 101–112
, doi = 10.1016/S0019-9958(59)90348-1
, issue = 2
, doi-access = free
Information theory
Probabilistic inequalities
Statistical inequalities