Klein's Inequality
   HOME

TheInfoList



OR:

In
mathematics Mathematics is an area of knowledge that includes the topics of numbers, formulas and related structures, shapes and the spaces in which they are contained, and quantities and their changes. These topics are represented in modern mathematics ...
, there are many kinds of inequalities involving matrices and
linear operator In mathematics, and more specifically in linear algebra, a linear map (also called a linear mapping, linear transformation, vector space homomorphism, or in some contexts linear function) is a mapping V \to W between two vector spaces that pre ...
s on
Hilbert space In mathematics, Hilbert spaces (named after David Hilbert) allow generalizing the methods of linear algebra and calculus from (finite-dimensional) Euclidean vector spaces to spaces that may be infinite-dimensional. Hilbert spaces arise natural ...
s. This article covers some important operator inequalities connected with traces of matrices.E. Carlen, Trace Inequalities and Quantum Entropy: An Introductory Course, Contemp. Math. 529 (2010) 73–140 B. Simon, Trace Ideals and their Applications, Cambridge Univ. Press, (1979); Second edition. Amer. Math. Soc., Providence, RI, (2005).


Basic definitions

Let H''n'' denote the space of Hermitian × matrices, H''n''+ denote the set consisting of positive semi-definite × Hermitian matrices and H''n''++ denote the set of
positive definite In mathematics, positive definiteness is a property of any object to which a bilinear form or a sesquilinear form may be naturally associated, which is positive-definite. See, in particular: * Positive-definite bilinear form * Positive-definite f ...
Hermitian matrices. For operators on an infinite dimensional Hilbert space we require that they be trace class and
self-adjoint In mathematics, and more specifically in abstract algebra, an element ''x'' of a *-algebra is self-adjoint if x^*=x. A self-adjoint element is also Hermitian, though the reverse doesn't necessarily hold. A collection ''C'' of elements of a sta ...
, in which case similar definitions apply, but we discuss only matrices, for simplicity. For any real-valued function on an interval ⊂ ℝ, one may define a matrix function for any operator with
eigenvalues In linear algebra, an eigenvector () or characteristic vector of a linear transformation is a nonzero vector that changes at most by a scalar factor when that linear transformation is applied to it. The corresponding eigenvalue, often denoted b ...
in by defining it on the eigenvalues and corresponding
projectors A projector or image projector is an optical device that projects an image (or moving images) onto a surface, commonly a projection screen. Most projectors create an image by shining a light through a small transparent lens, but some newer types ...
as :f(A)\equiv \sum_j f(\lambda_j)P_j ~, given the spectral decomposition A=\sum_j\lambda_j P_j.


Operator monotone

A function defined on an interval ⊂ ℝ is said to be operator monotone if ∀, and all with eigenvalues in , the following holds, :A \geq B \Rightarrow f(A) \geq f(B), where the inequality means that the operator is positive semi-definite. One may check that is, in fact, ''not'' operator monotone!


Operator convex

A function f: I \rightarrow \mathbb is said to be operator convex if for all n and all with eigenvalues in , and 0 < \lambda < 1, the following holds : f(\lambda A + (1-\lambda)B) \leq \lambda f(A) + (1 -\lambda)f(B) . Note that the operator \lambda A + (1-\lambda)B has eigenvalues in I, since A and B have eigenvalues in . A function f is operator concave if -f is operator convex, i.e. the inequality above for f is reversed.


Joint convexity

A function g: I\times J \rightarrow \mathbb, defined on intervals I,J\subset \mathbb is said to be jointly convex if for all n and all A_1, A_2\in \mathbf_n with eigenvalues in I and all B_1,B_2\in \mathbf_n with eigenvalues in J, and any 0\leq \lambda\leq 1 the following holds : g(\lambda A_1 + (1-\lambda)A_2,\lambda B_1 + (1-\lambda)B_2 ) \leq \lambda g(A_1, B_1) + (1 -\lambda)g(A_2, B_2). A function is jointly concave if − is jointly convex, i.e. the inequality above for is reversed.


Trace function

Given a function : ℝ → ℝ, the associated trace function on H''n'' is given by : A\mapsto \operatorname f(A)=\sum_j f(\lambda_j), where has eigenvalues and Tr stands for a
trace Trace may refer to: Arts and entertainment Music * Trace (Son Volt album), ''Trace'' (Son Volt album), 1995 * Trace (Died Pretty album), ''Trace'' (Died Pretty album), 1993 * Trace (band), a Dutch progressive rock band * The Trace (album), ''The ...
of the operator.


Convexity and monotonicity of the trace function

Let : ℝ → ℝ be continuous, and let be any integer. Then, if t\mapsto f(t) is monotone increasing, so is A \mapsto \operatorname f(A) on H''n''. Likewise, if t \mapsto f(t) is convex, so is A \mapsto \operatorname f(A) on H''n'', and it is strictly convex if is strictly convex. See proof and discussion in, for example.


Löwner–Heinz theorem

For -1\leq p \leq 0, the function f(t) = -t^p is operator monotone and operator concave. For 0 \leq p \leq 1, the function f(t) = t^p is operator monotone and operator concave. For 1 \leq p \leq 2, the function f(t) = t^p is operator convex. Furthermore, :f(t) = \log(t) is operator concave and operator monotone, while :f(t) = t \log(t) is operator convex. The original proof of this theorem is due to K. Löwner who gave a necessary and sufficient condition for to be operator monotone. An elementary proof of the theorem is discussed in and a more general version of it in.


Klein's inequality

For all Hermitian × matrices and and all differentiable
convex function In mathematics, a real-valued function is called convex if the line segment between any two points on the graph of a function, graph of the function lies above the graph between the two points. Equivalently, a function is convex if its epigra ...
s : ℝ → ℝ with
derivative In mathematics, the derivative of a function of a real variable measures the sensitivity to change of the function value (output value) with respect to a change in its argument (input value). Derivatives are a fundamental tool of calculus. F ...
, or for all positive-definite Hermitian × matrices and , and all differentiable convex functions :(0,∞) → ℝ, the following inequality holds, In either case, if is strictly convex, equality holds if and only if = . A popular choice in applications is , see below.


Proof

Let C=A-B so that, for t\in (0,1), :B + tC = (1 -t)B + tA, varies from B to A. Define :F(t) = \operatorname
(B + tC) B, or b, is the second Letter (alphabet), letter of the Latin-script alphabet, used in the English alphabet, modern English alphabet, the alphabets of other western European languages and others worldwide. Its name in English is ''English ...
/math>. By convexity and monotonicity of trace functions, F(t) is convex, and so for all t\in (0,1), : F(0) + t(F(1)-F(0))\geq F(t) , which is, : F(1) - F(0) \geq \frac , and, in fact, the right hand side is monotone decreasing in t. Taking the limit t\to 0 yields, : F(1) - F(0) \geq F'(0) , which with rearrangement and substitution is Klein's inequality: : \mathrm (A)-f(B)-(A-B)f'(B)\geq 0 Note that if f(t) is strictly convex and C\neq 0 , then F(t) is strictly convex. The final assertion follows from this and the fact that \tfrac is monotone decreasing in t.


Golden–Thompson inequality

In 1965, S. Golden and C.J. Thompson independently discovered that For any matrices A, B\in\mathbf_n, :\operatorname e^\leq \operatorname e^A e^B. This inequality can be generalized for three operators: for non-negative operators A, B, C\in\mathbf_n^+, :\operatorname e^\leq \int_0^\infty \operatorname A(B+t)^C(B+t)^\,\operatornamet.


Peierls–Bogoliubov inequality

Let R, F\in \mathbf_n be such that Tr e''R'' = 1. Defining , we have :\operatorname e^F e^R \geq \operatorname e^\geq e^g. The proof of this inequality follows from the above combined with
Klein's inequality In mathematics, there are many kinds of inequality (mathematics), inequalities involving matrix (mathematics), matrices and linear operators on Hilbert spaces. This article covers some important operator inequalities connected with Trace (linear alg ...
. Take .D. Ruelle, Statistical Mechanics: Rigorous Results, World Scient. (1969).


Gibbs variational principle

Let H be a self-adjoint operator such that e^ is trace class. Then for any \gamma\geq 0 with \operatorname\gamma=1, :\operatorname\gamma H+\operatorname\gamma\ln\gamma\geq -\ln \operatorname e^, with equality if and only if \gamma=\exp(-H)/\operatorname \exp(-H).


Lieb's concavity theorem

The following theorem was proved by E. H. Lieb in. It proves and generalizes a conjecture of E. P. Wigner, M. M. Yanase and F. J. Dyson. Six years later other proofs were given by T. Ando and B. Simon, and several more have been given since then. For all m\times n matrices K, and all q and r such that 0 \leq q\leq 1 and 0\leq r \leq 1, with q + r \leq 1 the real valued map on \mathbf^+_m \times \mathbf^+_n given by : F(A,B,K) = \operatorname(K^*A^qKB^r) * is jointly concave in (A,B) * is convex in K. Here K^* stands for the adjoint operator of K.


Lieb's theorem

For a fixed Hermitian matrix L\in\mathbf_n, the function : f(A)=\operatorname \exp\ is concave on \mathbf_n^. The theorem and proof are due to E. H. Lieb, Thm 6, where he obtains this theorem as a corollary of Lieb's concavity Theorem. The most direct proof is due to H. Epstein; see M.B. Ruskai papers, for a review of this argument.


Ando's convexity theorem

T. Ando's proof of
Lieb's concavity theorem In mathematics, there are many kinds of inequalities involving matrices and linear operators on Hilbert spaces. This article covers some important operator inequalities connected with traces of matrices.E. Carlen, Trace Inequalities and Quantum Entr ...
led to the following significant complement to it: For all m \times n matrices K, and all 1 \leq q \leq 2 and 0 \leq r \leq 1 with q-r \geq 1, the real valued map on \mathbf^_m \times \mathbf^_n given by : (A,B) \mapsto \operatorname(K^*A^qKB^) is convex.


Joint convexity of relative entropy

For two operators A, B\in\mathbf^_n define the following map : R(A\parallel B):= \operatorname(A\log A) - \operatorname(A\log B). For
density matrices In quantum mechanics, a density matrix (or density operator) is a matrix that describes the quantum state of a physical system. It allows for the calculation of the probabilities of the outcomes of any Measurement in quantum mechanics, measurement ...
\rho and \sigma, the map R(\rho\parallel\sigma)=S(\rho\parallel\sigma) is the Umegaki's
quantum relative entropy In quantum information theory, quantum relative entropy is a measure of distinguishability between two density matrix, quantum states. It is the quantum mechanical analog of relative entropy. Motivation For simplicity, it will be assumed that al ...
. Note that the non-negativity of R(A\parallel B) follows from Klein's inequality with f(t)=t\log t.


Statement

The map R(A\parallel B): \mathbf^_n \times \mathbf^_n \rightarrow \mathbf is jointly convex.


Proof

For all 0 < p < 1, (A,B) \mapsto \operatorname(B^A^p) is jointly concave, by
Lieb's concavity theorem In mathematics, there are many kinds of inequalities involving matrices and linear operators on Hilbert spaces. This article covers some important operator inequalities connected with traces of matrices.E. Carlen, Trace Inequalities and Quantum Entr ...
, and thus :(A,B)\mapsto \frac(\operatorname(B^A^p)-\operatornameA) is convex. But :\lim_\frac(\operatorname(B^A^p)-\operatornameA)=R(A\parallel B), and convexity is preserved in the limit. The proof is due to G. Lindblad.


Jensen's operator and trace inequalities

The operator version of
Jensen's inequality In mathematics, Jensen's inequality, named after the Danish mathematician Johan Jensen, relates the value of a convex function of an integral to the integral of the convex function. It was proved by Jensen in 1906, building on an earlier pr ...
is due to C. Davis.C. Davis, A Schwarz inequality for convex operator functions, Proc. Amer. Math. Soc. 8, 42–44, (1957). A continuous, real function f on an interval I satisfies Jensen's Operator Inequality if the following holds : f\left(\sum_kA_k^*X_kA_k\right)\leq\sum_k A_k^*f(X_k)A_k, for operators \_k with \sum_k A^*_kA_k=1 and for self-adjoint operators \_k with
spectrum A spectrum (plural ''spectra'' or ''spectrums'') is a condition that is not limited to a specific set of values but can vary, without gaps, across a continuum. The word was first used scientifically in optics to describe the rainbow of colors i ...
on I. See, for the proof of the following two theorems.


Jensen's trace inequality

Let be a continuous function defined on an interval and let and be natural numbers. If is convex, we then have the inequality : \operatorname\Bigl(f\Bigl(\sum_^nA_k^*X_kA_k\Bigr)\Bigr)\leq \operatorname\Bigl(\sum_^n A_k^*f(X_k)A_k\Bigr), for all (1, ... , ''n'') self-adjoint × matrices with spectra contained in and all (1, ... , ''n'') of × matrices with :\sum_^nA_k^*A_k=1. Conversely, if the above inequality is satisfied for some and , where > 1, then is convex.


Jensen's operator inequality

For a continuous function f defined on an interval I the following conditions are equivalent: * f is operator convex. * For each natural number n we have the inequality : f\Bigl(\sum_^nA_k^*X_kA_k\Bigr)\leq\sum_^n A_k^*f(X_k)A_k, for all (X_1, \ldots , X_n) bounded, self-adjoint operators on an arbitrary
Hilbert space In mathematics, Hilbert spaces (named after David Hilbert) allow generalizing the methods of linear algebra and calculus from (finite-dimensional) Euclidean vector spaces to spaces that may be infinite-dimensional. Hilbert spaces arise natural ...
\mathcal with spectra contained in I and all (A_1, \ldots , A_n) on \mathcal with \sum_^n A^*_kA_k=1. * f(V^*XV) \leq V^*f(X)V for each isometry V on an infinite-dimensional Hilbert space \mathcal and every self-adjoint operator X with spectrum in I. * Pf(PXP + \lambda(1 -P))P \leq Pf(X)P for each projection P on an infinite-dimensional Hilbert space \mathcal, every self-adjoint operator X with spectrum in I and every \lambda in I.


Araki–Lieb–Thirring inequality

E. H. Lieb and W. E. Thirring proved the following inequality in 1976: For any A\geq 0 , B\geq 0 and r\geq 1, :\operatorname ((BAB)^r)\leq \operatorname (B^A^B^). In 1990 H. Araki generalized the above inequality to the following one: For any A\geq 0 , B\geq 0 and q\geq 0, :\operatorname((BAB)^)\leq \operatorname((B^A^rB^)^q), for r\geq 1, and :\operatorname((B^A^rB^)^q)\leq \operatorname((BAB)^), for 0\leq r\leq 1. There are several other inequalities close to the Lieb–Thirring inequality, such as the following: for any A\geq 0 , B\geq 0 and \alpha \in ,1 :\operatorname (B A^\alpha B B A^ B) \leq \operatorname (B^2AB^2), and even more generally: for any A\geq 0 , B\geq 0 , r\geq 1/2 and c \geq 0, :\operatorname((B A B^ A B)^r) \leq \operatorname((B^A^2B^)^r). The above inequality generalizes the previous one, as can be seen by exchanging A by B^2 and B by A^ with \alpha = 2 c / (2 c + 2) and using the cyclicity of the trace, leading to :\operatorname((B A^ B B A^ B)^r) \leq \operatorname((B^ A B^2)^r).


Effros's theorem and its extension

E. Effros in proved the following theorem. If f(x) is an operator convex function, and L and R are commuting bounded linear operators, i.e. the commutator ,RLR-RL=0, the ''perspective'' :g(L, R):=f(LR^)R is jointly convex, i.e. if L=\lambda L_1+(1-\lambda)L_2 and R=\lambda R_1+(1-\lambda)R_2 with _i, R_i0 (i=1,2), 0\leq\lambda\leq 1, :g(L,R)\leq \lambda g(L_1,R_1)+(1-\lambda)g(L_2,R_2). Ebadian et al. later extended the inequality to the case where L and R do not commute .


Von Neumann's trace inequality and related results

Von Neumann's trace inequality, named after its originator
John von Neumann John von Neumann (; hu, Neumann János Lajos, ; December 28, 1903 – February 8, 1957) was a Hungarian-American mathematician, physicist, computer scientist, engineer and polymath. He was regarded as having perhaps the widest cove ...
, states that for any ''n'' × ''n'' complex matrices ''A'', ''B'' with singular values \alpha_1 \ge \alpha_2 \ge \cdots \ge \alpha_n and \beta_1 \ge \beta_2 \ge \cdots \ge \beta_n respectively, :\left, \operatorname (AB) \ \le \sum_^n \alpha_i \beta_i\,, with equality if and only if ''A'' and ''B'' share singular vectors. A simple corollary to this is the following result: For hermitian ''n'' × ''n'' positive semidefinite complex matrices ''A'', ''B'' where now the
eigenvalue In linear algebra, an eigenvector () or characteristic vector of a linear transformation is a nonzero vector that changes at most by a scalar factor when that linear transformation is applied to it. The corresponding eigenvalue, often denoted b ...
s are sorted decreasingly ( a_1 \ge a_2 \ge \cdots \ge a_n and b_1 \ge b_2 \ge \cdots \ge b_n, respectively), : \sum_^n a_i b_\leq \operatorname(AB)\leq \sum_^n a_i b_i\,.


See also

* * * *


References

{{reflist
Scholarpedia
primary source. Operator theory Matrix theory Inequalities