Diagonalizable matrix
   HOME

TheInfoList



OR:

In
linear algebra Linear algebra is the branch of mathematics concerning linear equations such as: :a_1x_1+\cdots +a_nx_n=b, linear maps such as: :(x_1, \ldots, x_n) \mapsto a_1x_1+\cdots +a_nx_n, and their representations in vector spaces and through matrices ...
, a square matrix A is called diagonalizable or non-defective if it is similar to a
diagonal matrix In linear algebra, a diagonal matrix is a matrix in which the entries outside the main diagonal are all zero; the term usually refers to square matrices. Elements of the main diagonal can either be zero or nonzero. An example of a 2×2 diagonal m ...
, i.e., if there exists an
invertible matrix In linear algebra, an -by- square matrix is called invertible (also nonsingular or nondegenerate), if there exists an -by- square matrix such that :\mathbf = \mathbf = \mathbf_n \ where denotes the -by- identity matrix and the multiplicati ...
P and a diagonal matrix D such that or equivalently (Such D are not unique.) For a finite-dimensional
vector space In mathematics and physics, a vector space (also called a linear space) is a set whose elements, often called '' vectors'', may be added together and multiplied ("scaled") by numbers called ''scalars''. Scalars are often real numbers, but can ...
a linear map T:V\to V is called diagonalizable if there exists an
ordered basis In mathematics, a set of vectors in a vector space is called a basis if every element of may be written in a unique way as a finite linear combination of elements of . The coefficients of this linear combination are referred to as components ...
of V consisting of
eigenvector In linear algebra, an eigenvector () or characteristic vector of a linear transformation is a nonzero vector that changes at most by a scalar factor when that linear transformation is applied to it. The corresponding eigenvalue, often denoted ...
s of T. These definitions are equivalent: if T has a matrix representation T = PDP^ as above, then the column vectors of P form a basis consisting of eigenvectors of and the diagonal entries of D are the corresponding
eigenvalue In linear algebra, an eigenvector () or characteristic vector of a linear transformation is a nonzero vector that changes at most by a scalar factor when that linear transformation is applied to it. The corresponding eigenvalue, often denoted ...
s of with respect to this eigenvector basis, A is represented by Diagonalization is the process of finding the above P and Diagonalizable matrices and maps are especially easy for computations, once their eigenvalues and eigenvectors are known. One can raise a diagonal matrix D to a power by simply raising the diagonal entries to that power, and the
determinant In mathematics, the determinant is a scalar value that is a function of the entries of a square matrix. It characterizes some properties of the matrix and the linear map represented by the matrix. In particular, the determinant is nonzero if a ...
of a diagonal matrix is simply the product of all diagonal entries; such computations generalize easily to Geometrically, a diagonalizable matrix is an inhomogeneous dilation (or ''anisotropic scaling'') — it scales the space, as does a '' homogeneous dilation'', but by a different factor along each eigenvector axis, the factor given by the corresponding eigenvalue. A square matrix that is not diagonalizable is called '' defective''. It can happen that a matrix A with
real Real may refer to: Currencies * Brazilian real (R$) * Central American Republic real * Mexican real * Portuguese real * Spanish real * Spanish colonial real Music Albums * ''Real'' (L'Arc-en-Ciel album) (2000) * ''Real'' (Bright album) (2010) ...
entries is defective over the real numbers, meaning that A = PDP^ is impossible for any invertible P and diagonal D with real entries, but it is possible with
complex Complex commonly refers to: * Complexity, the behaviour of a system whose components interact in multiple ways so possible interactions are difficult to describe ** Complex system, a system composed of many components which may interact with each ...
entries, so that A is diagonalizable over the complex numbers. For example, this is the case for a generic
rotation matrix In linear algebra, a rotation matrix is a transformation matrix that is used to perform a rotation in Euclidean space. For example, using the convention below, the matrix :R = \begin \cos \theta & -\sin \theta \\ \sin \theta & \cos \theta \en ...
. Many results for diagonalizable matrices hold only over an algebraically closed field (such as the complex numbers). In this case, diagonalizable matrices are
dense Density (volumetric mass density or specific mass) is the substance's mass per unit of volume. The symbol most often used for density is ''ρ'' (the lower case Greek letter rho), although the Latin letter ''D'' can also be used. Mathematically ...
in the space of all matrices, which means any defective matrix can be deformed into a diagonalizable matrix by a small perturbation; and the
Jordan normal form In linear algebra, a Jordan normal form, also known as a Jordan canonical form (JCF), is an upper triangular matrix of a particular form called a Jordan matrix representing a linear operator on a finite-dimensional vector space with respect to so ...
theorem states that any matrix is uniquely the sum of a diagonalizable matrix and a
nilpotent matrix In linear algebra, a nilpotent matrix is a square matrix ''N'' such that :N^k = 0\, for some positive integer k. The smallest such k is called the index of N, sometimes the degree of N. More generally, a nilpotent transformation is a linear tr ...
. Over an algebraically closed field, diagonalizable matrices are equivalent to semi-simple matrices.


Definition

A square n \times n matrix, A, with entries in a field F is called diagonalizable or nondefective if there exists an n \times n invertible matrix (i.e. an element of the
general linear group In mathematics, the general linear group of degree ''n'' is the set of invertible matrices, together with the operation of ordinary matrix multiplication. This forms a group, because the product of two invertible matrices is again invertible, ...
GL''n''(''F'')), P, such that P^AP is a diagonal matrix. Formally,


Characterization

The fundamental fact about diagonalizable maps and matrices is expressed by the following: * An n \times n matrix A over a field F is diagonalizable
if and only if In logic and related fields such as mathematics and philosophy, "if and only if" (shortened as "iff") is a biconditional logical connective between statements, where either both statements are true or both are false. The connective is b ...
the sum of the
dimension In physics and mathematics, the dimension of a mathematical space (or object) is informally defined as the minimum number of coordinates needed to specify any point within it. Thus, a line has a dimension of one (1D) because only one coor ...
s of its eigenspaces is equal to n, which is the case if and only if there exists a
basis Basis may refer to: Finance and accounting * Adjusted basis, the net cost of an asset after adjusting for various tax-related items *Basis point, 0.01%, often used in the context of interest rates * Basis trading, a trading strategy consisting ...
of F^ consisting of eigenvectors of A. If such a basis has been found, one can form the matrix P having these basis vectors as columns, and P^AP will be a diagonal matrix whose diagonal entries are the eigenvalues of A. The matrix P is known as a
modal matrix In linear algebra, the modal matrix is used in the diagonalization process involving eigenvalues and eigenvectors. Specifically the modal matrix M for the matrix A is the ''n'' × ''n'' matrix formed with the eigenvectors of A as columns in M. It ...
for A. * A linear map T : V \to V is diagonalizable if and only if the sum of the
dimension In physics and mathematics, the dimension of a mathematical space (or object) is informally defined as the minimum number of coordinates needed to specify any point within it. Thus, a line has a dimension of one (1D) because only one coor ...
s of its eigenspaces is equal to which is the case if and only if there exists a basis of V consisting of eigenvectors of T. With respect to such a basis, T will be represented by a diagonal matrix. The diagonal entries of this matrix are the eigenvalues of Another characterization: A matrix or linear map is diagonalizable over the field F if and only if its minimal polynomial is a product of distinct linear factors over (Put another way, a matrix is diagonalizable if and only if all of its elementary divisors are linear.) The following sufficient (but not necessary) condition is often useful. * An n \times n matrix A is diagonalizable over the field F if it has n distinct eigenvalues in i.e. if its characteristic polynomial has n distinct roots in however, the converse may be false. Consider \begin -1 & 3 & -1 \\ -3 & 5 & -1 \\ -3 & 3 & 1 \end, which has eigenvalues 1, 2, 2 (not all distinct) and is diagonalizable with diagonal form ( similar to \begin 1 & 0 & 0 \\ 0 & 2 & 0 \\ 0 & 0 & 2 \end and change of basis matrix P: \begin 1 & 1 & -1 \\ 1 & 1 & 0 \\ 1 & 0 & 3 \end. The converse fails when A has an eigenspace of dimension higher than 1. In this example, the eigenspace of A associated with the eigenvalue 2 has dimension 2. * A linear map T : V \to V with n = \dim(V) is diagonalizable if it has n distinct eigenvalues, i.e. if its characteristic polynomial has n distinct roots in F. Let A be a matrix over If A is diagonalizable, then so is any power of it. Conversely, if A is invertible, F is algebraically closed, and A^n is diagonalizable for some n that is not an integer multiple of the characteristic of then A is diagonalizable. Proof: If A^n is diagonalizable, then A is annihilated by some polynomial which has no multiple root (since and is divided by the minimal polynomial of Over the complex numbers \Complex, almost every matrix is diagonalizable. More precisely: the set of complex n \times n matrices that are ''not'' diagonalizable over considered as a subset of has Lebesgue measure zero. One can also say that the diagonalizable matrices form a dense subset with respect to the
Zariski topology In algebraic geometry and commutative algebra, the Zariski topology is a topology which is primarily defined by its closed sets. It is very different from topologies which are commonly used in the real or complex analysis; in particular, it is n ...
: the non-diagonalizable matrices lie inside the vanishing set of the discriminant of the characteristic polynomial, which is a
hypersurface In geometry, a hypersurface is a generalization of the concepts of hyperplane, plane curve, and surface. A hypersurface is a manifold or an algebraic variety of dimension , which is embedded in an ambient space of dimension , generally a Euclidea ...
. From that follows also density in the usual (''strong'') topology given by a norm. The same is not true over The
Jordan–Chevalley decomposition In mathematics, the Jordan–Chevalley decomposition, named after Camille Jordan and Claude Chevalley, expresses a linear operator as the sum of its commuting semisimple part and its nilpotent part. The multiplicative decomposition expresses an inve ...
expresses an operator as the sum of its semisimple (i.e., diagonalizable) part and its nilpotent part. Hence, a matrix is diagonalizable if and only if its nilpotent part is zero. Put in another way, a matrix is diagonalizable if each block in its
Jordan form In linear algebra, a Jordan normal form, also known as a Jordan canonical form (JCF), is an upper triangular matrix of a particular form called a Jordan matrix representing a linear operator on a finite-dimensional vector space with respect to so ...
has no nilpotent part; i.e., each "block" is a one-by-one matrix.


Diagonalization

If a matrix A can be diagonalized, that is, : P^AP = \begin \lambda_1 & 0 & \cdots & 0 \\ 0 & \lambda_2 & \cdots & 0 \\ \vdots & \vdots & \ddots & \vdots \\ 0 & 0 & \cdots & \lambda_n \end, then: : AP = P\begin \lambda_1 & 0 & \cdots & 0 \\ 0 & \lambda_2 & \cdots & 0 \\ \vdots & \vdots & \ddots & \vdots \\ 0 & 0 & \cdots & \lambda_n \end. Writing P as a block matrix of its column vectors \boldsymbol_ :P = \begin \boldsymbol_1 & \boldsymbol_2 & \cdots & \boldsymbol_n \end, the above equation can be rewritten as :A\boldsymbol_i = \lambda_i \boldsymbol_i \qquad (i=1,2,\dots,n). So the column vectors of P are
right eigenvector In linear algebra, an eigenvector () or characteristic vector of a linear transformation is a nonzero vector that changes at most by a scalar factor when that linear transformation is applied to it. The corresponding eigenvalue, often denoted ...
s of and the corresponding diagonal entry is the corresponding
eigenvalue In linear algebra, an eigenvector () or characteristic vector of a linear transformation is a nonzero vector that changes at most by a scalar factor when that linear transformation is applied to it. The corresponding eigenvalue, often denoted ...
. The invertibility of P also suggests that the eigenvectors are
linearly independent In the theory of vector spaces, a set of vectors is said to be if there is a nontrivial linear combination of the vectors that equals the zero vector. If no such linear combination exists, then the vectors are said to be . These concepts are ...
and form a basis of This is the necessary and sufficient condition for diagonalizability and the canonical approach of diagonalization. The
row vector In linear algebra, a column vector with m elements is an m \times 1 matrix consisting of a single column of m entries, for example, \boldsymbol = \begin x_1 \\ x_2 \\ \vdots \\ x_m \end. Similarly, a row vector is a 1 \times n matrix for some n, c ...
s of P^ are the
left eigenvector In linear algebra, an eigenvector () or characteristic vector of a linear transformation is a nonzero vector that changes at most by a scalar factor when that linear transformation is applied to it. The corresponding eigenvalue, often denoted ...
s of When a complex matrix A\in\mathbb^ is a
Hermitian matrix In mathematics, a Hermitian matrix (or self-adjoint matrix) is a complex square matrix that is equal to its own conjugate transpose—that is, the element in the -th row and -th column is equal to the complex conjugate of the element in the -t ...
(or more generally a
normal matrix In mathematics, a complex square matrix is normal if it commutes with its conjugate transpose : The concept of normal matrices can be extended to normal operators on infinite dimensional normed spaces and to normal elements in C*-algebras. As ...
), eigenvectors of A can be chosen to form an orthonormal basis of and P can be chosen to be a
unitary matrix In linear algebra, a complex square matrix is unitary if its conjugate transpose is also its inverse, that is, if U^* U = UU^* = UU^ = I, where is the identity matrix. In physics, especially in quantum mechanics, the conjugate transpose is ...
. If in addition, A\in\mathbb^ is a real
symmetric matrix In linear algebra, a symmetric matrix is a square matrix that is equal to its transpose. Formally, Because equal matrices have equal dimensions, only square matrices can be symmetric. The entries of a symmetric matrix are symmetric with ...
, then its eigenvectors can be chosen to be an orthonormal basis of \mathbb^n and P can be chosen to be an
orthogonal matrix In linear algebra, an orthogonal matrix, or orthonormal matrix, is a real square matrix whose columns and rows are orthonormal vectors. One way to express this is Q^\mathrm Q = Q Q^\mathrm = I, where is the transpose of and is the identity m ...
. For most practical work matrices are diagonalized numerically using computer software. Many algorithms exist to accomplish this.


Simultaneous diagonalization

A set of matrices is said to be ''simultaneously diagonalizable'' if there exists a single invertible matrix P such that P^AP is a diagonal matrix for every A in the set. The following theorem characterizes simultaneously diagonalizable matrices: A set of diagonalizable matrices commutes if and only if the set is simultaneously diagonalizable. The set of all n \times n diagonalizable matrices (over with n > 1 is not simultaneously diagonalizable. For instance, the matrices : \begin 1 & 0 \\ 0 & 0 \end \quad\text\quad \begin 1 & 1 \\ 0 & 0 \end are diagonalizable but not simultaneously diagonalizable because they do not commute. A set consists of commuting normal matrices if and only if it is simultaneously diagonalizable by a
unitary matrix In linear algebra, a complex square matrix is unitary if its conjugate transpose is also its inverse, that is, if U^* U = UU^* = UU^ = I, where is the identity matrix. In physics, especially in quantum mechanics, the conjugate transpose is ...
; that is, there exists a unitary matrix U such that U^ AU is diagonal for every A in the set. In the language of
Lie theory In mathematics, the mathematician Sophus Lie ( ) initiated lines of study involving integration of differential equations, transformation groups, and contact of spheres that have come to be called Lie theory. For instance, the latter subject is ...
, a set of simultaneously diagonalizable matrices generate a toral Lie algebra.


Examples


Diagonalizable matrices

*
Involution Involution may refer to: * Involute, a construction in the differential geometry of curves * '' Agricultural Involution: The Processes of Ecological Change in Indonesia'', a 1963 study of intensification of production through increased labour inpu ...
s are diagonalizable over the reals (and indeed any field of characteristic not 2), with ±1 on the diagonal. * Finite order
endomorphism In mathematics, an endomorphism is a morphism from a mathematical object to itself. An endomorphism that is also an isomorphism is an automorphism. For example, an endomorphism of a vector space is a linear map , and an endomorphism of a gr ...
s are diagonalizable over \mathbb (or any algebraically closed field where the characteristic of the field does not divide the order of the endomorphism) with
roots of unity In mathematics, a root of unity, occasionally called a de Moivre number, is any complex number that yields 1 when raised to some positive integer power . Roots of unity are used in many branches of mathematics, and are especially important in ...
on the diagonal. This follows since the minimal polynomial is separable, because the roots of unity are distinct. * Projections are diagonalizable, with 0s and 1s on the diagonal. * Real
symmetric matrices In linear algebra, a symmetric matrix is a square matrix that is equal to its transpose. Formally, Because equal matrices have equal dimensions, only square matrices can be symmetric. The entries of a symmetric matrix are symmetric with re ...
are diagonalizable by orthogonal matrices; i.e., given a real symmetric matrix Q^AQ is diagonal for some orthogonal matrix More generally, matrices are diagonalizable by unitary matrices if and only if they are
normal Normal(s) or The Normal(s) may refer to: Film and television * ''Normal'' (2003 film), starring Jessica Lange and Tom Wilkinson * ''Normal'' (2007 film), starring Carrie-Anne Moss, Kevin Zegers, Callum Keith Rennie, and Andrew Airlie * ''Norma ...
. In the case of the real symmetric matrix, we see that so clearly AA^ = A^A holds. Examples of normal matrices are real symmetric (or skew-symmetric) matrices (e.g. covariance matrices) and
Hermitian matrices In mathematics, a Hermitian matrix (or self-adjoint matrix) is a complex square matrix that is equal to its own conjugate transpose—that is, the element in the -th row and -th column is equal to the complex conjugate of the element in the -th ...
(or skew-Hermitian matrices). See
spectral theorem In mathematics, particularly linear algebra and functional analysis, a spectral theorem is a result about when a linear operator or matrix can be diagonalized (that is, represented as a diagonal matrix in some basis). This is extremely useful ...
s for generalizations to infinite-dimensional vector spaces.


Matrices that are not diagonalizable

In general, a
rotation matrix In linear algebra, a rotation matrix is a transformation matrix that is used to perform a rotation in Euclidean space. For example, using the convention below, the matrix :R = \begin \cos \theta & -\sin \theta \\ \sin \theta & \cos \theta \en ...
is not diagonalizable over the reals, but all rotation matrices are diagonalizable over the complex field. Even if a matrix is not diagonalizable, it is always possible to "do the best one can", and find a matrix with the same properties consisting of eigenvalues on the leading diagonal, and either ones or zeroes on the superdiagonal – known as
Jordan normal form In linear algebra, a Jordan normal form, also known as a Jordan canonical form (JCF), is an upper triangular matrix of a particular form called a Jordan matrix representing a linear operator on a finite-dimensional vector space with respect to so ...
. Some matrices are not diagonalizable over any field, most notably nonzero nilpotent matrices. This happens more generally if the algebraic and geometric multiplicities of an eigenvalue do not coincide. For instance, consider : C = \begin 0 & 1 \\ 0 & 0 \end. This matrix is not diagonalizable: there is no matrix U such that U^CU is a diagonal matrix. Indeed, C has one eigenvalue (namely zero) and this eigenvalue has algebraic multiplicity 2 and geometric multiplicity 1. Some real matrices are not diagonalizable over the reals. Consider for instance the matrix : B = \left begin 0 & 1 \\ \!-1 & 0 \end\right The matrix B does not have any real eigenvalues, so there is no real matrix Q such that Q^BQ is a diagonal matrix. However, we can diagonalize B if we allow complex numbers. Indeed, if we take : Q = \begin 1 & i \\ i & 1 \end, then Q^BQ is diagonal. It is easy to find that B is the rotation matrix which rotates counterclockwise by angle \theta = \frac Note that the above examples show that the sum of diagonalizable matrices need not be diagonalizable.


How to diagonalize a matrix

Diagonalizing a matrix is the same process as finding its
eigenvalues and eigenvectors In linear algebra, an eigenvector () or characteristic vector of a linear transformation is a nonzero vector that changes at most by a scalar factor when that linear transformation is applied to it. The corresponding eigenvalue, often denote ...
, in the case that the eigenvectors form a basis. For example, consider the matrix :A=\left begin 0 & 1 & \!\!\!-2\\ 0 & 1 & 0\\ 1 & \!\!\!-1 & 3 \end\right The roots of the characteristic polynomial p(\lambda)=\det(\lambda I-A) are the eigenvalues Solving the linear system \left(I-A\right) \mathbf = \mathbf gives the eigenvectors \mathbf_1 = (1,1,0) and while \left(2I-A\right)\mathbf = \mathbf gives that is, A \mathbf_i = \lambda_i \mathbf_i for These vectors form a basis of so we can assemble them as the column vectors of a change-of-basis matrix P to get: P^AP = \left begin 1 & 0 & 1\\ 1 & 2 & 0\\ 0 & 1 & \!\!\!\!-1 \end\right \left begin 0 & 1 & \!\!\!-2\\ 0 & 1 & 0\\ 1 & \!\!\!-1 & 3 \end\right \left begin 1 & \,0 & 1\\ 1 & 2 & 0\\ 0 & 1 & \!\!\!\!-1 \end\right= \begin 1 & 0 & 0 \\ 0 & 1 & 0 \\ 0 & 0 & 2 \end = D . We may see this equation in terms of transformations: P takes the standard basis to the eigenbasis, so we have: P^ AP \mathbf_i = P^ A \mathbf_i = P^ (\lambda_i\mathbf_i) = \lambda_i\mathbf_i, so that P^ AP has the standard basis as its eigenvectors, which is the defining property of Note that there is no preferred order of the eigenvectors in changing the order of the
eigenvectors In linear algebra, an eigenvector () or characteristic vector of a linear transformation is a nonzero vector that changes at most by a scalar factor when that linear transformation is applied to it. The corresponding eigenvalue, often denoted ...
in P just changes the order of the eigenvalues in the diagonalized form of


Application to matrix functions

Diagonalization can be used to efficiently compute the powers of a matrix : \begin A^k &= \left(PDP^\right)^k = \left(PDP^\right) \left(PDP^\right) \cdots \left(PDP^\right) \\ &= PD\left(P^P\right) D \left(P^P\right) \cdots \left(P^P\right) D P^ = PD^kP^, \end and the latter is easy to calculate since it only involves the powers of a diagonal matrix. For example, for the matrix A with eigenvalues \lambda = 1,1,2 in the example above we compute: : \begin A^k = PD^kP^ &= \left begin 1 & \,0 & 1 \\ 1 & 2 & 0 \\ 0 & 1 & \!\!\!\!-1 \end\right \begin 1^k & 0 & 0 \\ 0 & 1^k & 0 \\ 0 & 0 & 2^k \end \left begin 1 & \,0 & 1 \\ 1 & 2 & 0 \\ 0 & 1 & \!\!\!\!-1 \end\right \\ em &= \begin 2 - 2^k & -1 + 2^k & 2 - 2^ \\ 0 & 1 & 0 \\ -1 + 2^k & 1 - 2^k & -1 + 2^ \end. \end This approach can be generalized to
matrix exponential In mathematics, the matrix exponential is a matrix function on square matrices analogous to the ordinary exponential function. It is used to solve systems of linear differential equations. In the theory of Lie groups, the matrix exponential give ...
and other
matrix function In mathematics, every analytic function can be used for defining a matrix function that maps square matrices with complex entries to square matrices of the same size. This is used for defining the exponential of a matrix, which is involved in th ...
s that can be defined as power series. For example, defining we have: : \begin \exp(A) = P \exp(D) P^ &= \left begin 1 & \,0 & 1 \\ 1 & 2 & 0 \\ 0 & 1 & \!\!\!\!-1 \end\right \begin e^1 & 0 & 0 \\ 0 & e^1 & 0 \\ 0 & 0 & e^2 \end \left begin 1 & \,0 & 1\\ 1 & 2 & 0\\ 0 & 1 & \!\!\!\!-1 \end\right \\ em &= \begin 2 e - e^2 & -e + e^2 & 2 e - 2 e^2 \\ 0 & e & 0 \\ -e + e^2 & e - e^2 & -e + 2 e^2 \end. \end This is particularly useful in finding closed form expressions for terms of linear recursive sequences, such as the
Fibonacci numbers In mathematics, the Fibonacci numbers, commonly denoted , form a sequence, the Fibonacci sequence, in which each number is the sum of the two preceding ones. The sequence commonly starts from 0 and 1, although some authors start the sequence from ...
.


Particular application

For example, consider the following matrix: :M = \begina & b - a\\ 0 & b\end. Calculating the various powers of M reveals a surprising pattern: : M^2 = \begina^2 & b^2-a^2 \\ 0 &b^2 \end,\quad M^3 = \begina^3 & b^3-a^3 \\ 0 &b^3 \end,\quad M^4 = \begina^4 & b^4-a^4 \\ 0 &b^4 \end,\quad \ldots The above phenomenon can be explained by diagonalizing To accomplish this, we need a basis of \R^2 consisting of eigenvectors of One such eigenvector basis is given by : \mathbf = \begin 1 \\ 0 \end = \mathbf_1,\quad \mathbf = \begin 1 \\ 1 \end = \mathbf_1 + \mathbf_2, where e''i'' denotes the standard basis of R''n''. The reverse change of basis is given by :\mathbf_1 = \mathbf,\qquad \mathbf_2 = \mathbf - \mathbf. Straightforward calculations show that :M\mathbf = a\mathbf,\qquad M\mathbf = b\mathbf. Thus, ''a'' and ''b'' are the eigenvalues corresponding to u and v, respectively. By linearity of matrix multiplication, we have that : M^n \mathbf = a^n \mathbf,\qquad M^n \mathbf = b^n \mathbf. Switching back to the standard basis, we have :\begin M^n \mathbf_1 &= M^n \mathbf = a^n \mathbf_1, \\ M^n \mathbf_2 &= M^n \left(\mathbf - \mathbf\right) = b^n \mathbf - a^n\mathbf = \left(b^n - a^n\right) \mathbf_1 + b^n\mathbf_2. \end The preceding relations, expressed in matrix form, are :M^n = \begin a^n & b^n - a^n \\ 0 & b^n \end, thereby explaining the above phenomenon.


Quantum mechanical application

In
quantum mechanical Quantum mechanics is a fundamental theory in physics that provides a description of the physical properties of nature at the scale of atoms and subatomic particles. It is the foundation of all quantum physics including quantum chemistry, qua ...
and
quantum chemical Quantum chemistry, also called molecular quantum mechanics, is a branch of physical chemistry focused on the application of quantum mechanics to chemical systems, particularly towards the quantum-mechanical calculation of electronic contributions ...
computations matrix diagonalization is one of the most frequently applied numerical processes. The basic reason is that the time-independent
Schrödinger equation The Schrödinger equation is a linear partial differential equation that governs the wave function of a quantum-mechanical system. It is a key result in quantum mechanics, and its discovery was a significant landmark in the development of th ...
is an eigenvalue equation, albeit in most of the physical situations on an infinite dimensional space (a Hilbert space). A very common approximation is to truncate Hilbert space to finite dimension, after which the Schrödinger equation can be formulated as an eigenvalue problem of a real symmetric, or complex Hermitian matrix. Formally this approximation is founded on the
variational principle In science and especially in mathematical studies, a variational principle is one that enables a problem to be solved using calculus of variations, which concerns finding functions that optimize the values of quantities that depend on those funct ...
, valid for Hamiltonians that are bounded from below. First-order perturbation theory also leads to matrix eigenvalue problem for degenerate states.


See also

* Defective matrix *
Scaling (geometry) In affine geometry, uniform scaling (or isotropic scaling) is a linear transformation that enlarges (increases) or shrinks (diminishes) objects by a '' scale factor'' that is the same in all directions. The result of uniform scaling is simil ...
*
Triangular matrix In mathematics, a triangular matrix is a special kind of square matrix. A square matrix is called if all the entries ''above'' the main diagonal are zero. Similarly, a square matrix is called if all the entries ''below'' the main diagonal are ...
*
Semisimple operator In mathematics, a linear operator ''T'' on a vector space is semisimple if every ''T''-invariant subspace has a complementary ''T''-invariant subspace;Lam (2001), p. 39/ref> in other words, the vector space is a semisimple representation of the o ...
* Diagonalizable group *
Jordan normal form In linear algebra, a Jordan normal form, also known as a Jordan canonical form (JCF), is an upper triangular matrix of a particular form called a Jordan matrix representing a linear operator on a finite-dimensional vector space with respect to so ...
* Weight module – associative algebra generalization *
Orthogonal diagonalization In linear algebra, an orthogonal diagonalization of a symmetric matrix is a diagonalization by means of an orthogonal change of coordinates. The following is an orthogonal diagonalization algorithm that diagonalizes a quadratic form ''q''(''x'') o ...


Notes


References

{{Matrix classes Matrices