Differential entropy (also referred to as continuous entropy) is a concept in
information theory
Information theory is the mathematical study of the quantification (science), quantification, Data storage, storage, and telecommunications, communication of information. The field was established and formalized by Claude Shannon in the 1940s, ...
that began as an attempt by
Claude Shannon
Claude Elwood Shannon (April 30, 1916 – February 24, 2001) was an American mathematician, electrical engineer, computer scientist, cryptographer and inventor known as the "father of information theory" and the man who laid the foundations of th ...
to extend the idea of (Shannon)
entropy
Entropy is a scientific concept, most commonly associated with states of disorder, randomness, or uncertainty. The term and the concept are used in diverse fields, from classical thermodynamics, where it was first recognized, to the micros ...
(a measure of average
surprisal) of a
random variable
A random variable (also called random quantity, aleatory variable, or stochastic variable) is a Mathematics, mathematical formalization of a quantity or object which depends on randomness, random events. The term 'random variable' in its mathema ...
, to continuous probability distributions. Unfortunately, Shannon did not derive this formula, and rather just assumed it was the correct continuous analogue of discrete entropy, but it is not. The actual continuous version of discrete entropy is the
limiting density of discrete points (LDDP). Differential entropy (described here) is commonly encountered in the literature, but it is a limiting case of the LDDP, and one that loses its fundamental association with discrete
entropy
Entropy is a scientific concept, most commonly associated with states of disorder, randomness, or uncertainty. The term and the concept are used in diverse fields, from classical thermodynamics, where it was first recognized, to the micros ...
.
In terms of
measure theory
In mathematics, the concept of a measure is a generalization and formalization of geometrical measures (length, area, volume) and other common notions, such as magnitude (mathematics), magnitude, mass, and probability of events. These seemingl ...
, the differential entropy of a
probability measure
In mathematics, a probability measure is a real-valued function defined on a set of events in a σ-algebra that satisfies Measure (mathematics), measure properties such as ''countable additivity''. The difference between a probability measure an ...
is the negative
relative entropy
Relative may refer to:
General use
*Kinship and family, the principle binding the most basic social units of society. If two people are connected by circumstances of birth, they are said to be ''relatives''.
Philosophy
*Relativism, the concept t ...
from that measure to the
Lebesgue measure
In measure theory, a branch of mathematics, the Lebesgue measure, named after French mathematician Henri Lebesgue, is the standard way of assigning a measure to subsets of higher dimensional Euclidean '-spaces. For lower dimensions or , it c ...
, where the latter is treated as if it were a probability measure, despite being unnormalized.
Definition
Let
be a random variable with a
probability density function
In probability theory, a probability density function (PDF), density function, or density of an absolutely continuous random variable, is a Function (mathematics), function whose value at any given sample (or point) in the sample space (the s ...
whose
support is a set
. The ''differential entropy''
or
is defined as
For probability distributions which do not have an explicit density function expression, but have an explicit
quantile function expression,
, then
can be defined in terms of the derivative of
i.e. the quantile density function
as
As with its discrete analog, the units of differential entropy depend on the base of the
logarithm
In mathematics, the logarithm of a number is the exponent by which another fixed value, the base, must be raised to produce that number. For example, the logarithm of to base is , because is to the rd power: . More generally, if , the ...
, which is usually 2 (i.e., the units are
bits). See
logarithmic units for logarithms taken in different bases. Related concepts such as
joint
A joint or articulation (or articular surface) is the connection made between bones, ossicles, or other hard structures in the body which link an animal's skeletal system into a functional whole.Saladin, Ken. Anatomy & Physiology. 7th ed. McGraw- ...
,
conditional differential entropy, and
relative entropy
Relative may refer to:
General use
*Kinship and family, the principle binding the most basic social units of society. If two people are connected by circumstances of birth, they are said to be ''relatives''.
Philosophy
*Relativism, the concept t ...
are defined in a similar fashion. Unlike the discrete analog, the differential entropy has an offset that depends on the units used to measure
.
For example, the differential entropy of a quantity measured in millimeters will be more than the same quantity measured in meters; a dimensionless quantity will have differential entropy of more than the same quantity divided by 1000.
One must take care in trying to apply properties of discrete entropy to differential entropy, since probability density functions can be greater than 1. For example, the
uniform distribution has ''negative'' differential entropy; i.e., it is better ordered than
as shown now
being less than that of
which has ''zero'' differential entropy. Thus, differential entropy does not share all properties of discrete entropy.
The continuous
mutual information
In probability theory and information theory, the mutual information (MI) of two random variables is a measure of the mutual Statistical dependence, dependence between the two variables. More specifically, it quantifies the "Information conten ...
has the distinction of retaining its fundamental significance as a measure of discrete information since it is actually the limit of the discrete mutual information of ''partitions'' of
and
as these partitions become finer and finer. Thus it is invariant under non-linear
homeomorphisms (continuous and uniquely invertible maps), including linear
transformations of
and
, and still represents the amount of discrete information that can be transmitted over a channel that admits a continuous space of values.
For the direct analogue of discrete entropy extended to the continuous space, see
limiting density of discrete points.
Properties of differential entropy
* For probability densities
and
, the
Kullback–Leibler divergence
In mathematical statistics, the Kullback–Leibler (KL) divergence (also called relative entropy and I-divergence), denoted D_\text(P \parallel Q), is a type of statistical distance: a measure of how much a model probability distribution is diff ...
is greater than or equal to 0 with equality only if
almost everywhere
In measure theory (a branch of mathematical analysis), a property holds almost everywhere if, in a technical sense, the set for which the property holds takes up nearly all possibilities. The notion of "almost everywhere" is a companion notion to ...
. Similarly, for two random variables
and
,
and
with equality
if and only if
In logic and related fields such as mathematics and philosophy, "if and only if" (often shortened as "iff") is paraphrased by the biconditional, a logical connective between statements. The biconditional is true in two cases, where either bo ...
and
are
independent
Independent or Independents may refer to:
Arts, entertainment, and media Artist groups
* Independents (artist group), a group of modernist painters based in Pennsylvania, United States
* Independentes (English: Independents), a Portuguese artist ...
.
* The chain rule for differential entropy holds as in the discrete case
* Differential entropy is translation invariant, i.e. for a constant
.
* Differential entropy is in general not invariant under arbitrary invertible maps.In particular, for a constant
,
For a vector valued random variable
and an invertible (square)
matrix
Matrix (: matrices or matrixes) or MATRIX may refer to:
Science and mathematics
* Matrix (mathematics), a rectangular array of numbers, symbols or expressions
* Matrix (logic), part of a formula in prenex normal form
* Matrix (biology), the m ...
* In general, for a transformation from a random vector to another random vector with same dimension
, the corresponding entropies are related via
where
is the
Jacobian of the transformation
. The above inequality becomes an equality if the transform is a bijection. Furthermore, when
is a rigid rotation, translation, or combination thereof, the Jacobian determinant is always 1, and
.
* If a random vector
has mean zero and
covariance
In probability theory and statistics, covariance is a measure of the joint variability of two random variables.
The sign of the covariance, therefore, shows the tendency in the linear relationship between the variables. If greater values of one ...
matrix
,
, ,
\begin
&\log \frac B \\ pt&+ \left(1 - \frac\right) \psi \\ pt&- \left(1 + \frac\right)\psi \\ pt&+ \frac \psi
\end, ,
">-
, Gamma , , f(x) = \frac , , \log(\theta \Gamma(k)) + \left(1 - k\right) \psi(k) + k , ,
">-
, Laplace distribution, Laplace , , f(x) = \frac \exp\left(-\frac\right) , , 1 + \log(2b) \, , , (-\infty,\infty)\,
, -
, Logistic , , f(x) = \frac, , \log s + 2 \, , , (-\infty,\infty)\,
, -
, Log-normal distribution, Lognormal , , f(x) = \frac \exp\left(-\frac\right) , , \mu + \tfrac \log(2\pi e \sigma^2), , [0,\infty)\,
, -
, Maxwell–Boltzmann distribution, Maxwell–Boltzmann , , f(x) = \frac\sqrt\,x^\exp\left(-\frac\right) , , \log(a\sqrt) + \gamma_E - \tfrac, , [0,\infty)\,
, -
, Generalized Gaussian distribution, Generalized normal , , f(x) = \frac x^ \exp\left(-\beta x^2\right), , \log - \frac \psi\left(\frac\right) + \frac, , (-\infty,\infty)\,
, -
, Pareto , , f(x) = \frac , , \log \frac + 1 + \frac, ,
">-
, Student's t , , f(x) = \frac , , \begin
&\frac \left \
&+\!\log \sqrt B
\end">, (-\infty,\infty)\,
, -
, Triangular distribution, Triangular , , f(x) = \begin
\frac & \mathrm a \le x \leq c, \\ pt \frac & \mathrm c < x \le b, \\ pt \end , , \frac + \log \frac, , [a,b]\,
, -
, Weibull distribution, Weibull , , f(x) = \frac x^ \exp\left(-\frac\right) , , \frac\gamma_E + \log \frac + 1, , [0,\infty)\,
, -
, Multivariate normal distribution, Multivariate normal , ,
f_X(\mathbf) = \frac , , \tfrac \log\left 2\pi e)^N \det(\Sigma)\right/math>, , \mathbb^N
Many of the differential entropies are from.
Variants
As described above, differential entropy does not share all properties of discrete entropy. For example, the differential entropy can be negative; also it is not invariant under continuous coordinate transformations. Edwin Thompson Jaynes
Edwin Thompson Jaynes (July 5, 1922 – April 30, 1998) was the Wayman Crow Distinguished Professor of Physics at Washington University in St. Louis. He wrote extensively on statistical mechanics and on foundations of probability and statistic ...
showed in fact that the expression above is not the correct limit of the expression for a finite set of probabilities.
A modification of differential entropy adds an invariant measure
In mathematics, an invariant measure is a measure that is preserved by some function. The function may be a geometric transformation. For examples, circular angle is invariant under rotation, hyperbolic angle is invariant under squeeze mappin ...
factor to correct this, (see limiting density of discrete points). If m(x) is further constrained to be a probability density, the resulting notion is called relative entropy
Relative may refer to:
General use
*Kinship and family, the principle binding the most basic social units of society. If two people are connected by circumstances of birth, they are said to be ''relatives''.
Philosophy
*Relativism, the concept t ...
in information theory:
D(p\parallel m) = \int p(x)\log\frac\,dx.
The definition of differential entropy above can be obtained by partitioning the range of X into bins of length h with associated sample points ih within the bins, for X Riemann integrable. This gives a quantized version of X, defined by X_h = ih if ih \le X \le (i+1)h. Then the entropy of X_h = ih is
H_h=-\sum_i hf(ih)\log (f(ih)) - \sum hf(ih)\log(h).
The first term on the right approximates the differential entropy, while the second term is approximately -\log(h). Note that this procedure suggests that the entropy in the discrete sense of a continuous random variable
In probability theory and statistics, a probability distribution is a function that gives the probabilities of occurrence of possible events for an experiment. It is a mathematical description of a random phenomenon in terms of its sample spa ...
should be \infty.
See also
*Information entropy
In information theory, the entropy of a random variable quantifies the average level of uncertainty or information associated with the variable's potential states or possible outcomes. This measures the expected amount of information needed ...
* Self-information
* Entropy estimation
References
External links
*
* {{planetmath reference, urlname=DifferentialEntropy, title=Differential entropy
Entropy and information
Information theory
Statistical randomness