In
quantum information theory
Quantum information is the information of the state of a quantum system. It is the basic entity of study in quantum information theory, and can be manipulated using quantum information processing techniques. Quantum information refers to both ...
, quantum relative entropy is a measure of distinguishability between two
quantum states. It is the quantum mechanical analog of
relative entropy.
Motivation
For simplicity, it will be assumed that all objects in the article are finite-dimensional.
We first discuss the classical case. Suppose the probabilities of a finite sequence of events is given by the probability distribution ''P'' = , but somehow we mistakenly assumed it to be ''Q'' = . For instance, we can mistake an unfair coin for a fair one. According to this erroneous assumption, our uncertainty about the ''j''-th event, or equivalently, the amount of information provided after observing the ''j''-th event, is
:
The (assumed) average uncertainty of all possible events is then
:
On the other hand, the
Shannon entropy of the probability distribution ''p'', defined by
:
is the real amount of uncertainty before observation. Therefore the difference between these two quantities
:
is a measure of the distinguishability of the two probability distributions ''p'' and ''q''. This is precisely the classical relative entropy, or
Kullback–Leibler divergence
In mathematical statistics, the Kullback–Leibler divergence (also called relative entropy and I-divergence), denoted D_\text(P \parallel Q), is a type of statistical distance: a measure of how one probability distribution ''P'' is different fro ...
:
:
Note
#In the definitions above, the convention that 0·log 0 = 0 is assumed, since
. Intuitively, one would expect that an event of
zero probability to contribute nothing towards entropy.
#The relative entropy is not a
metric. For example, it is not symmetric. The uncertainty discrepancy in mistaking a fair coin to be unfair is not the same as the opposite situation.
Definition
As with many other objects in quantum information theory, quantum relative entropy is defined by extending the classical definition from probability distributions to
density matrices
In quantum mechanics, a density matrix (or density operator) is a matrix that describes the quantum state of a physical system. It allows for the calculation of the probabilities of the outcomes of any Measurement in quantum mechanics, measurement ...
. Let ''ρ'' be a density matrix. The
von Neumann entropy of ''ρ'', which is the quantum mechanical analog of the Shannon entropy, is given by
:
For two density matrices ''ρ'' and ''σ'', the quantum relative entropy of ''ρ'' with respect to ''σ'' is defined by
:
We see that, when the states are classically related, i.e. ''ρσ'' = ''σρ'', the definition coincides with the classical case, in the sense that if
and
with
and
(because
and
commute, they are
simultaneously diagonalizable
In linear algebra, a square matrix A is called diagonalizable or non-defective if it is similar to a diagonal matrix, i.e., if there exists an invertible matrix P and a diagonal matrix D such that or equivalently (Such D are not unique.) ...
), then
is just the ordinary
Kullback-Leibler divergence of the probability vector
with respect to the probability vector
.
Non-finite (divergent) relative entropy
In general, the ''support'' of a matrix ''M'' is the orthogonal complement of its
kernel
Kernel may refer to:
Computing
* Kernel (operating system), the central component of most operating systems
* Kernel (image processing), a matrix used for image convolution
* Compute kernel, in GPGPU programming
* Kernel method, in machine lea ...
, i.e.
. When considering the quantum relative entropy, we assume the convention that −''s'' · log 0 = ∞ for any ''s'' > 0. This leads to the definition that
:
when
:
This can be interpreted in the following way. Informally, the quantum relative entropy is a measure of our ability to distinguish two quantum states where larger values indicate states that are more different. Being orthogonal represents the most different quantum states can be. This is reflected by non-finite quantum relative entropy for orthogonal quantum states. Following the argument given in the Motivation section, if we erroneously assume the state
has support in
, this is an error impossible to recover from.
However, one should be careful not to conclude that the divergence of the quantum relative entropy
implies that the states
and
are orthogonal or even very different by other measures. Specifically,
can diverge when
and
differ by a ''vanishingly small amount'' as measured by some norm. For example, let
have the diagonal representation
with
for
and
for
where
is an orthonormal set. The kernel of
is the space spanned by the set
. Next let
for a small positive number
. As
has support (namely the state
) in the kernel of
,
is divergent even though the trace norm of the difference
is
. This means that difference between
and
as measured by the trace norm is vanishingly small as
even though
is divergent (i.e. infinite). This property of the quantum relative entropy represents a serious shortcoming if not treated with care.
Non-negativity of relative entropy
Corresponding classical statement
For the classical
Kullback–Leibler divergence
In mathematical statistics, the Kullback–Leibler divergence (also called relative entropy and I-divergence), denoted D_\text(P \parallel Q), is a type of statistical distance: a measure of how one probability distribution ''P'' is different fro ...
, it can be shown that
:
and the equality holds if and only if ''P'' = ''Q''. Colloquially, this means that the uncertainty calculated using erroneous assumptions is always greater than the real amount of uncertainty.
To show the inequality, we rewrite
:
Notice that log is a
concave function. Therefore -log is
convex
Convex or convexity may refer to:
Science and technology
* Convex lens, in optics
Mathematics
* Convex set, containing the whole line segment that joins points
** Convex polygon, a polygon which encloses a convex set of points
** Convex polytop ...
. Applying
Jensen's inequality, we obtain
:
Jensen's inequality also states that equality holds if and only if, for all ''i'', ''q
i'' = (Σ''q
j'') ''p
i'', i.e. ''p'' = ''q''.
The result
Klein's inequality states that the quantum relative entropy
:
is non-negative in general. It is zero if and only if ''ρ'' = ''σ''.
Proof
Let ''ρ'' and ''σ'' have spectral decompositions
:
So
:
Direct calculation gives
:
:
:
where ''P
i j'' = , ''v
i*w
j'',
2.
Since the matrix (''P
i j'')''
i j'' is a
doubly stochastic matrix and -log is a convex function, the above expression is
:
Define ''r''
i = Σ
''j''''q
j P
i j''. Then is a probability distribution. From the non-negativity of classical relative entropy, we have
:
The second part of the claim follows from the fact that, since -log is strictly convex, equality is achieved in
:
if and only if (''P
i j'') is a
permutation matrix, which implies ''ρ'' = ''σ'', after a suitable labeling of the eigenvectors and .
Joint convexity of relative entropy
The relative entropy is
jointly convex. For
and states
we have
Monotonicity of relative entropy
The relative entropy decreases monotonically under
completely positive trace preserving (CPTP) operations
on density matrices,
.
This inequality is called Monotonicity of quantum relative entropy and was first proved by
Lindblad.
An entanglement measure
Let a composite quantum system have state space
:
and ''ρ'' be a density matrix acting on ''H''.
The relative entropy of entanglement of ''ρ'' is defined by
:
where the minimum is taken over the family of
separable states. A physical interpretation of the quantity is the optimal distinguishability of the state ''ρ'' from separable states.
Clearly, when ''ρ'' is not
entangled
:
by Klein's inequality.
Relation to other quantum information quantities
One reason the quantum relative entropy is useful is that several other important quantum information quantities are special cases of it. Often, theorems are stated in terms of the quantum relative entropy, which lead to immediate corollaries concerning the other quantities. Below, we list some of these relations.
Let ''ρ''
AB be the joint state of a bipartite system with subsystem ''A'' of dimension ''n''
A and ''B'' of dimension ''n''
B. Let ''ρ''
A, ''ρ''
B be the respective reduced states, and ''I''
A, ''I''
B the respective identities. The
maximally mixed states are ''I''
A/''n''
A and ''I''
B/''n''
B. Then it is possible to show with direct computation that
:
:
:
where ''I''(''A'':''B'') is the
quantum mutual information and ''S''(''B'', ''A'') is the
quantum conditional entropy.
References
* {{cite journal , last=Vedral , first=V. , title=The role of relative entropy in quantum information theory , journal=Reviews of Modern Physics , publisher=American Physical Society (APS) , volume=74 , issue=1 , date=8 March 2002 , issn=0034-6861 , doi=10.1103/revmodphys.74.197 , pages=197–234, arxiv=quant-ph/0102094, bibcode=2002RvMP...74..197V , s2cid=6370982
* Michael A. Nielsen, Isaac L. Chuang
"Quantum Computation and Quantum Information"* Marco Tomamichel,
Quantum Information Processing with Finite Resources -- Mathematical Foundations.
arXiv:1504.00233
Quantum mechanical entropy
Quantum information theory