HOME

TheInfoList



OR:

In
linear algebra Linear algebra is the branch of mathematics concerning linear equations such as: :a_1x_1+\cdots +a_nx_n=b, linear maps such as: :(x_1, \ldots, x_n) \mapsto a_1x_1+\cdots +a_nx_n, and their representations in vector spaces and through matrices ...
, eigendecomposition is the
factorization In mathematics, factorization (or factorisation, see English spelling differences) or factoring consists of writing a number or another mathematical object as a product of several ''factors'', usually smaller or simpler objects of the same kin ...
of a
matrix Matrix most commonly refers to: * ''The Matrix'' (franchise), an American media franchise ** ''The Matrix'', a 1999 science-fiction action film ** "The Matrix", a fictional setting, a virtual reality environment, within ''The Matrix'' (franchis ...
into a
canonical form In mathematics and computer science, a canonical, normal, or standard form of a mathematical object is a standard way of presenting that object as a mathematical expression. Often, it is one which provides the simplest representation of an obj ...
, whereby the matrix is represented in terms of its
eigenvalues and eigenvectors In linear algebra, an eigenvector () or characteristic vector of a linear transformation is a nonzero vector that changes at most by a scalar factor when that linear transformation is applied to it. The corresponding eigenvalue, often denoted ...
. Only
diagonalizable matrices In linear algebra, a square matrix A is called diagonalizable or non-defective if it is similar to a diagonal matrix, i.e., if there exists an invertible matrix P and a diagonal matrix D such that or equivalently (Such D are not unique.) F ...
can be factorized in this way. When the matrix being factorized is a normal or real
symmetric matrix In linear algebra, a symmetric matrix is a square matrix that is equal to its transpose. Formally, Because equal matrices have equal dimensions, only square matrices can be symmetric. The entries of a symmetric matrix are symmetric with r ...
, the decomposition is called "spectral decomposition", derived from the
spectral theorem In mathematics, particularly linear algebra and functional analysis, a spectral theorem is a result about when a linear operator or matrix can be diagonalized (that is, represented as a diagonal matrix in some basis). This is extremely useful be ...
.


Fundamental theory of matrix eigenvectors and eigenvalues

A (nonzero) vector of dimension is an eigenvector of a square matrix if it satisfies a linear equation of the form :\mathbf \mathbf = \lambda \mathbf for some scalar . Then is called the eigenvalue corresponding to . Geometrically speaking, the eigenvectors of are the vectors that merely elongates or shrinks, and the amount that they elongate/shrink by is the eigenvalue. The above equation is called the eigenvalue equation or the eigenvalue problem. This yields an equation for the eigenvalues : p\left(\lambda\right) = \det\left(\mathbf - \lambda \mathbf\right)= 0. We call the
characteristic polynomial In linear algebra, the characteristic polynomial of a square matrix is a polynomial which is invariant under matrix similarity and has the eigenvalues as roots. It has the determinant and the trace of the matrix among its coefficients. The char ...
, and the equation, called the characteristic equation, is an th order polynomial equation in the unknown . This equation will have distinct solutions, where . The set of solutions, that is, the eigenvalues, is called the
spectrum A spectrum (plural ''spectra'' or ''spectrums'') is a condition that is not limited to a specific set of values but can vary, without gaps, across a continuum. The word was first used scientifically in optics to describe the rainbow of colors ...
of . If the field of scalars is
algebraically closed In mathematics, a field is algebraically closed if every non-constant polynomial in (the univariate polynomial ring with coefficients in ) has a root in . Examples As an example, the field of real numbers is not algebraically closed, becau ...
, then we can factor as :p\left(\lambda\right) = \left(\lambda - \lambda_1\right)^\left(\lambda - \lambda_2\right)^ \cdots \left(\lambda-\lambda_\right)^ = 0. The integer is termed the
algebraic multiplicity In linear algebra, an eigenvector () or characteristic vector of a linear transformation is a nonzero vector that changes at most by a scalar factor when that linear transformation is applied to it. The corresponding eigenvalue, often denoted ...
of eigenvalue . The algebraic multiplicities sum to : \sum_^ = N. For each eigenvalue , we have a specific eigenvalue equation :\left(\mathbf - \lambda_i \mathbf\right)\mathbf = 0. There will be
linearly independent In the theory of vector spaces, a set of vectors is said to be if there is a nontrivial linear combination of the vectors that equals the zero vector. If no such linear combination exists, then the vectors are said to be . These concepts a ...
solutions to each eigenvalue equation. The linear combinations of the solutions (except the one which gives the zero vector) are the eigenvectors associated with the eigenvalue . The integer is termed the
geometric multiplicity In linear algebra, an eigenvector () or characteristic vector of a linear transformation is a nonzero vector that changes at most by a scalar factor when that linear transformation is applied to it. The corresponding eigenvalue, often denoted ...
of . It is important to keep in mind that the algebraic multiplicity and geometric multiplicity may or may not be equal, but we always have . The simplest case is of course when . The total number of linearly independent eigenvectors, , can be calculated by summing the geometric multiplicities :\sum_^ = N_. The eigenvectors can be indexed by eigenvalues, using a double index, with being the th eigenvector for the th eigenvalue. The eigenvectors can also be indexed using the simpler notation of a single index , with .


Eigendecomposition of a matrix

Let be a square matrix with linearly independent eigenvectors (where ). Then can be factorized as :\mathbf=\mathbf\mathbf\mathbf^ where is the square matrix whose th column is the eigenvector of , and is the
diagonal matrix In linear algebra, a diagonal matrix is a matrix in which the entries outside the main diagonal are all zero; the term usually refers to square matrices. Elements of the main diagonal can either be zero or nonzero. An example of a 2×2 diagonal m ...
whose diagonal elements are the corresponding eigenvalues, . Note that only
diagonalizable matrices In linear algebra, a square matrix A is called diagonalizable or non-defective if it is similar to a diagonal matrix, i.e., if there exists an invertible matrix P and a diagonal matrix D such that or equivalently (Such D are not unique.) F ...
can be factorized in this way. For example, the
defective matrix In linear algebra, a defective matrix is a square matrix that does not have a complete basis of eigenvectors, and is therefore not diagonalizable. In particular, an ''n'' × ''n'' matrix is defective if and only if it does not hav ...
\left \begin 1 & 1 \\ 0 & 1 \end \right/math> (which is a shear matrix) cannot be diagonalized. The eigenvectors are usually normalized, but they need not be. A non-normalized set of eigenvectors, can also be used as the columns of . That can be understood by noting that the magnitude of the eigenvectors in gets canceled in the decomposition by the presence of . If one of the eigenvalues has more than one linearly independent eigenvectors (that is, the geometric multiplicity of is greater than 1), then these eigenvectors for this eigenvalue can be chosen to be mutually orthogonal; however, if two eigenvectors belong to two different eigenvalues, it may be impossible for them to be orthogonal to each other (see Example below). One special case is that if is a normal matrix, then by the spectral theorem, it's always possible to diagonalize in an orthonormal basis . The decomposition can be derived from the fundamental property of eigenvectors: :\begin \mathbf \mathbf &= \lambda \mathbf \\ \mathbf \mathbf &= \mathbf \mathbf \\ \mathbf &= \mathbf\mathbf\mathbf^ . \end The linearly independent eigenvectors with nonzero eigenvalues form an orthogonal basis (not necessarily orthonormal) for all possible products , for , which is the same as the
image An image is a visual representation of something. It can be two-dimensional, three-dimensional, or somehow otherwise feed into the visual system to convey information. An image can be an artifact, such as a photograph or other two-dimensiona ...
(or range) of the corresponding
matrix transformation In linear algebra, linear transformations can be represented by matrices. If T is a linear transformation mapping \mathbb^n to \mathbb^m and \mathbf x is a column vector with n entries, then T( \mathbf x ) = A \mathbf x for some m \times n matrix ...
, and also the
column space In linear algebra, the column space (also called the range or image) of a matrix ''A'' is the span (set of all possible linear combinations) of its column vectors. The column space of a matrix is the image or range of the corresponding matr ...
of the matrix . The number of linearly independent eigenvectors with nonzero eigenvalues is equal to the
rank Rank is the relative position, value, worth, complexity, power, importance, authority, level, etc. of a person or object within a ranking, such as: Level or position in a hierarchical organization * Academic rank * Diplomatic rank * Hierarchy * H ...
of the matrix , and also the dimension of the image (or range) of the corresponding matrix transformation, as well as its column space. The linearly independent eigenvectors with an eigenvalue of zero form a basis (which can be chosen to be orthonormal) for the
null space In mathematics, the kernel of a linear map, also known as the null space or nullspace, is the linear subspace of the domain of the map which is mapped to the zero vector. That is, given a linear map between two vector spaces and , the kernel of ...
(also known as the kernel) of the matrix transformation .


Example

The 2 × 2 real matrix :\mathbf = \begin 1 & 0 \\ 1 & 3 \\ \end may be decomposed into a diagonal matrix through multiplication of a non-singular matrix :\mathbf = \begin a & b \\ c & d \end \in \mathbb^. Then : \begin a & b \\ c & d \end^\begin 1 & 0 \\ 1 & 3 \end\begin a & b \\ c & d \end = \begin x & 0 \\ 0 & y \end, for some real diagonal matrix \left \begin x & 0 \\ 0 & y \end \right/math>. Multiplying both sides of the equation on the left by : : \begin 1 & 0 \\ 1 & 3 \end \begin a & b \\ c & d \end = \begin a & b \\ c & d \end \begin x & 0 \\ 0 & y \end. The above equation can be decomposed into two
simultaneous equation In mathematics, a set of simultaneous equations, also known as a system of equations or an equation system, is a finite set of equations for which common solutions are sought. An equation system is usually classified in the same manner as single e ...
s: : \begin \begin 1 & 0\\ 1 & 3 \end \begin a \\ c \end = \begin ax \\ cx \end \\ \begin 1 & 0\\ 1 & 3 \end \begin b \\ d \end = \begin by \\ dy \end \end . Factoring out the
eigenvalue In linear algebra, an eigenvector () or characteristic vector of a linear transformation is a nonzero vector that changes at most by a scalar factor when that linear transformation is applied to it. The corresponding eigenvalue, often denoted ...
s and : : \begin \begin 1 & 0\\ 1 & 3 \end \begin a \\ c \end = x\begin a \\ c \end \\ \begin 1 & 0\\ 1 & 3 \end \begin b \\ d \end = y\begin b \\ d \end \end Letting :\mathbf = \begin a \\ c \end, \quad \mathbf = \begin b \\ d \end, this gives us two vector equations: : \begin \mathbf \mathbf = x \mathbf \\ \mathbf \mathbf = y \mathbf \end And can be represented by a single vector equation involving two solutions as eigenvalues: : \mathbf \mathbf = \lambda \mathbf where represents the two eigenvalues and , and represents the vectors and . Shifting to the left hand side and factoring out : (\mathbf - \lambda \mathbf) \mathbf = \mathbf Since is non-singular, it is essential that is nonzero. Therefore, : \det(\mathbf - \lambda \mathbf) = 0 Thus : (1- \lambda)(3 - \lambda) = 0 giving us the solutions of the eigenvalues for the matrix as or , and the resulting diagonal matrix from the eigendecomposition of is thus \left \begin 1 & 0 \\ 0 & 3 \end \right/math>. Putting the solutions back into the above simultaneous equations : \begin \begin 1 & 0 \\ 1 & 3 \end \begin a \\ c \end = 1\begin a \\ c \end \\ \begin 1 & 0\\ 1 & 3 \end \begin b \\ d \end = 3\begin b \\ d \end \end Solving the equations, we have :a = -2c \quad\text \quad b = 0, \qquad c,d \in \mathbb. Thus the matrix required for the eigendecomposition of is :\mathbf = \begin -2c & 0 \\ c & d \end,\qquad c, d\in \mathbb, that is: : \begin -2c & 0 \\ c & d \end^ \begin 1 & 0 \\ 1 & 3 \end \begin -2c & 0 \\ c & d \end = \begin 1 & 0 \\ 0 & 3 \end,\qquad c, d\in \mathbb


Matrix inverse via eigendecomposition

If a matrix can be eigendecomposed and if none of its eigenvalues are zero, then is
invertible In mathematics, the concept of an inverse element generalises the concepts of opposite () and reciprocal () of numbers. Given an operation denoted here , and an identity element denoted , if , one says that is a left inverse of , and that i ...
and its inverse is given by :\mathbf^ = \mathbf\mathbf^\mathbf^ If \mathbf is a symmetric matrix, since \mathbf is formed from the eigenvectors of \mathbf, \mathbf is guaranteed to be an
orthogonal matrix In linear algebra, an orthogonal matrix, or orthonormal matrix, is a real square matrix whose columns and rows are orthonormal vectors. One way to express this is Q^\mathrm Q = Q Q^\mathrm = I, where is the transpose of and is the identity ma ...
, therefore \mathbf^ = \mathbf^\mathrm. Furthermore, because is a
diagonal matrix In linear algebra, a diagonal matrix is a matrix in which the entries outside the main diagonal are all zero; the term usually refers to square matrices. Elements of the main diagonal can either be zero or nonzero. An example of a 2×2 diagonal m ...
, its inverse is easy to calculate: :\left Lambda^\right = \frac


Practical implications

When eigendecomposition is used on a matrix of measured, real
data In the pursuit of knowledge, data (; ) is a collection of discrete values that convey information, describing quantity, quality, fact, statistics, other basic units of meaning, or simply sequences of symbols that may be further interprete ...
, the inverse may be less valid when all eigenvalues are used unmodified in the form above. This is because as eigenvalues become relatively small, their contribution to the inversion is large. Those near zero or at the "noise" of the measurement system will have undue influence and could hamper solutions (detection) using the inverse. Two mitigations have been proposed: truncating small or zero eigenvalues, and extending the lowest reliable eigenvalue to those below it. The first mitigation method is similar to a sparse sample of the original matrix, removing components that are not considered valuable. However, if the solution or detection process is near the noise level, truncating may remove components that influence the desired solution. The second mitigation extends the eigenvalue so that lower values have much less influence over inversion, but do still contribute, such that solutions near the noise will still be found. The reliable eigenvalue can be found by assuming that eigenvalues of extremely similar and low value are a good representation of measurement noise (which is assumed low for most systems). If the eigenvalues are rank-sorted by value, then the reliable eigenvalue can be found by minimization of the
Laplacian In mathematics, the Laplace operator or Laplacian is a differential operator given by the divergence of the gradient of a scalar function on Euclidean space. It is usually denoted by the symbols \nabla\cdot\nabla, \nabla^2 (where \nabla is the ...
of the sorted eigenvalues: :\min\left, \nabla^2 \lambda_\mathrm\ where the eigenvalues are subscripted with an to denote being sorted. The position of the minimization is the lowest reliable eigenvalue. In measurement systems, the square root of this reliable eigenvalue is the average noise over the components of the system.


Functional calculus

The eigendecomposition allows for much easier computation of
power series In mathematics, a power series (in one variable) is an infinite series of the form \sum_^\infty a_n \left(x - c\right)^n = a_0 + a_1 (x - c) + a_2 (x - c)^2 + \dots where ''an'' represents the coefficient of the ''n''th term and ''c'' is a const ...
of matrices. If is given by :f(x) = a_0 + a_1 x + a_2 x^2 + \cdots then we know that :f\!\left(\mathbf\right) = \mathbf\,f\!\left(\mathbf\right)\mathbf^ Because is a
diagonal matrix In linear algebra, a diagonal matrix is a matrix in which the entries outside the main diagonal are all zero; the term usually refers to square matrices. Elements of the main diagonal can either be zero or nonzero. An example of a 2×2 diagonal m ...
, functions of are very easy to calculate: :\left \left(\mathbf\right)\right = f\left(\lambda_i\right) The off-diagonal elements of are zero; that is, is also a diagonal matrix. Therefore, calculating reduces to just calculating the function on each of the eigenvalues. A similar technique works more generally with the
holomorphic functional calculus In mathematics, holomorphic functional calculus is functional calculus with holomorphic functions. That is to say, given a holomorphic function ''f'' of a complex argument ''z'' and an operator ''T'', the aim is to construct an operator, ''f''('' ...
, using :\mathbf^ = \mathbf\mathbf^\mathbf^ from above. Once again, we find that :\left \left(\mathbf\right)\right = f\left(\lambda_i\right)


Examples

:\begin \mathbf^2 &= \left(\mathbf\mathbf\mathbf^\right)\left(\mathbf\mathbf\mathbf^\right) = \mathbf\mathbf\left(\mathbf^\mathbf\right)\mathbf\mathbf^ = \mathbf\mathbf^2\mathbf^ \\ \mathbf^n &= \mathbf\mathbf^n\mathbf^ \\ \exp \mathbf &= \mathbf \exp(\mathbf) \mathbf^ \end which are examples for the functions f(x)=x^2, \; f(x)=x^n, \; f(x)=\exp . Furthermore, \exp is the
matrix exponential In mathematics, the matrix exponential is a matrix function on square matrices analogous to the ordinary exponential function. It is used to solve systems of linear differential equations. In the theory of Lie groups, the matrix exponential gives ...
.


Decomposition for special matrices

When is normal or real symmetric matrix, the decomposition is called "spectral decomposition", derived from the spectral theorem.


Normal matrices

A complex-valued square matrix is normal (meaning , where is the conjugate transpose) if and only if it can be decomposed as :\mathbf = \mathbf\mathbf\mathbf^* where is a
unitary matrix In linear algebra, a complex square matrix is unitary if its conjugate transpose is also its inverse, that is, if U^* U = UU^* = UU^ = I, where is the identity matrix. In physics, especially in quantum mechanics, the conjugate transpose ...
(meaning ) and is a
diagonal matrix In linear algebra, a diagonal matrix is a matrix in which the entries outside the main diagonal are all zero; the term usually refers to square matrices. Elements of the main diagonal can either be zero or nonzero. An example of a 2×2 diagonal m ...
. The columns u1, ..., u''n'' of form an
orthonormal basis In mathematics, particularly linear algebra, an orthonormal basis for an inner product space ''V'' with finite dimension is a basis for V whose vectors are orthonormal, that is, they are all unit vectors and orthogonal to each other. For exam ...
and are eigenvectors of with corresponding eigenvalues ''λ''1, ..., ''λ''''n''. If is restricted to be a
Hermitian matrix In mathematics, a Hermitian matrix (or self-adjoint matrix) is a complex square matrix that is equal to its own conjugate transpose—that is, the element in the -th row and -th column is equal to the complex conjugate of the element in the ...
(), then has only real valued entries. If is restricted to a unitary matrix, then takes all its values on the complex unit circle, that is, .


Real symmetric matrices

As a special case, for every real symmetric matrix, the eigenvalues are real and the eigenvectors can be chosen real and
orthonormal In linear algebra, two vectors in an inner product space are orthonormal if they are orthogonal (or perpendicular along a line) unit vectors. A set of vectors form an orthonormal set if all vectors in the set are mutually orthogonal and all of uni ...
. Thus a real symmetric matrix can be decomposed as :\mathbf = \mathbf\mathbf\mathbf^\mathsf where is an
orthogonal matrix In linear algebra, an orthogonal matrix, or orthonormal matrix, is a real square matrix whose columns and rows are orthonormal vectors. One way to express this is Q^\mathrm Q = Q Q^\mathrm = I, where is the transpose of and is the identity ma ...
whose columns are (the above chosen, real and orthonormal) eigenvectors of , and is a diagonal matrix whose entries are the eigenvalues of .


Useful facts


Useful facts regarding eigenvalues

*The product of the eigenvalues is equal to the
determinant In mathematics, the determinant is a scalar value that is a function of the entries of a square matrix. It characterizes some properties of the matrix and the linear map represented by the matrix. In particular, the determinant is nonzero if an ...
of \det\left(\mathbf\right) = \prod_^ Note that each eigenvalue is raised to the power , the
algebraic multiplicity In linear algebra, an eigenvector () or characteristic vector of a linear transformation is a nonzero vector that changes at most by a scalar factor when that linear transformation is applied to it. The corresponding eigenvalue, often denoted ...
. *The sum of the eigenvalues is equal to the
trace Trace may refer to: Arts and entertainment Music * ''Trace'' (Son Volt album), 1995 * ''Trace'' (Died Pretty album), 1993 * Trace (band), a Dutch progressive rock band * ''The Trace'' (album) Other uses in arts and entertainment * ''Trace'' ...
of \operatorname\left(\mathbf\right) = \sum_^ Note that each eigenvalue is multiplied by , the
algebraic multiplicity In linear algebra, an eigenvector () or characteristic vector of a linear transformation is a nonzero vector that changes at most by a scalar factor when that linear transformation is applied to it. The corresponding eigenvalue, often denoted ...
. *If the eigenvalues of are , and is invertible, then the eigenvalues of are simply . *If the eigenvalues of are , then the eigenvalues of are simply , for any
holomorphic function In mathematics, a holomorphic function is a complex-valued function of one or more complex variables that is complex differentiable in a neighbourhood of each point in a domain in complex coordinate space . The existence of a complex deriv ...
.


Useful facts regarding eigenvectors

* If is
Hermitian {{Short description, none Numerous things are named after the French mathematician Charles Hermite (1822–1901): Hermite * Cubic Hermite spline, a type of third-degree spline * Gauss–Hermite quadrature, an extension of Gaussian quadrature me ...
and full-rank, the basis of eigenvectors may be chosen to be mutually
orthogonal In mathematics, orthogonality is the generalization of the geometric notion of ''perpendicularity''. By extension, orthogonality is also used to refer to the separation of specific features of a system. The term also has specialized meanings in ...
. The eigenvalues are real. * The eigenvectors of are the same as the eigenvectors of . * Eigenvectors are only defined up to a multiplicative constant. That is, if then is also an eigenvector for any scalar . In particular, and (for any ''θ'') are also eigenvectors. * In the case of degenerate eigenvalues (an eigenvalue having more than one eigenvector), the eigenvectors have an additional freedom of linear transformation, that is to say, any linear (orthonormal) combination of eigenvectors sharing an eigenvalue (in the degenerate subspace) is itself an eigenvector (in the subspace).


Useful facts regarding eigendecomposition

* can be eigendecomposed if and only if the number of linearly independent eigenvectors, , equals the dimension of an eigenvector: * If the field of scalars is algebraically closed and if has no repeated roots, that is, if N_\lambda = N, then can be eigendecomposed. * The statement " can be eigendecomposed" does ''not'' imply that has an inverse. * The statement " has an inverse" does ''not'' imply that can be eigendecomposed. A counterexample is \left \begin 1 & 1 \\ 0 & 1 \end \right/math>, which is an invertible
defective matrix In linear algebra, a defective matrix is a square matrix that does not have a complete basis of eigenvectors, and is therefore not diagonalizable. In particular, an ''n'' × ''n'' matrix is defective if and only if it does not hav ...
.


Useful facts regarding matrix inverse

* can be inverted
if and only if In logic and related fields such as mathematics and philosophy, "if and only if" (shortened as "iff") is a biconditional logical connective between statements, where either both statements are true or both are false. The connective is bic ...
all eigenvalues are nonzero: \lambda_i \ne 0 \quad \forall \,i * If ''and'' , the inverse is given by \mathbf^ = \mathbf\mathbf^\mathbf^


Numerical computations


Numerical computation of eigenvalues

Suppose that we want to compute the eigenvalues of a given matrix. If the matrix is small, we can compute them symbolically using the
characteristic polynomial In linear algebra, the characteristic polynomial of a square matrix is a polynomial which is invariant under matrix similarity and has the eigenvalues as roots. It has the determinant and the trace of the matrix among its coefficients. The char ...
. However, this is often impossible for larger matrices, in which case we must use a
numerical method In numerical analysis, a numerical method is a mathematical tool designed to solve numerical problems. The implementation of a numerical method with an appropriate convergence check in a programming language is called a numerical algorithm. Mathem ...
. In practice, eigenvalues of large matrices are not computed using the characteristic polynomial. Computing the polynomial becomes expensive in itself, and exact (symbolic) roots of a high-degree polynomial can be difficult to compute and express: the
Abel–Ruffini theorem In mathematics, the Abel–Ruffini theorem (also known as Abel's impossibility theorem) states that there is no solution in radicals to general polynomial equations of degree five or higher with arbitrary coefficients. Here, ''general'' means t ...
implies that the roots of high-degree (5 or above) polynomials cannot in general be expressed simply using th roots. Therefore, general algorithms to find eigenvectors and eigenvalues are
iterative Iteration is the repetition of a process in order to generate a (possibly unbounded) sequence of outcomes. Each repetition of the process is a single iteration, and the outcome of each iteration is then the starting point of the next iteration. ...
. Iterative numerical algorithms for approximating roots of polynomials exist, such as
Newton's method In numerical analysis, Newton's method, also known as the Newton–Raphson method, named after Isaac Newton and Joseph Raphson, is a root-finding algorithm which produces successively better approximations to the roots (or zeroes) of a real- ...
, but in general it is impractical to compute the characteristic polynomial and then apply these methods. One reason is that small
round-off error A roundoff error, also called rounding error, is the difference between the result produced by a given algorithm using exact arithmetic and the result produced by the same algorithm using finite-precision, rounded arithmetic. Rounding errors are d ...
s in the coefficients of the characteristic polynomial can lead to large errors in the eigenvalues and eigenvectors: the roots are an extremely
ill-conditioned In numerical analysis, the condition number of a function measures how much the output value of the function can change for a small change in the input argument. This is used to measure how sensitive a function is to changes or errors in the input ...
function of the coefficients. A simple and accurate iterative method is the
power method In mathematics, power iteration (also known as the power method) is an eigenvalue algorithm: given a diagonalizable matrix A, the algorithm will produce a number \lambda, which is the greatest (in absolute value) eigenvalue of A, and a nonzero vec ...
: a
random In common usage, randomness is the apparent or actual lack of pattern or predictability in events. A random sequence of events, symbols or steps often has no order and does not follow an intelligible pattern or combination. Individual rand ...
vector is chosen and a sequence of
unit vector In mathematics, a unit vector in a normed vector space is a vector (often a spatial vector) of length 1. A unit vector is often denoted by a lowercase letter with a circumflex, or "hat", as in \hat (pronounced "v-hat"). The term ''direction vec ...
s is computed as : \frac, \frac, \frac, \ldots This
sequence In mathematics, a sequence is an enumerated collection of objects in which repetitions are allowed and order matters. Like a set, it contains members (also called ''elements'', or ''terms''). The number of elements (possibly infinite) is call ...
will
almost always In probability theory, an event is said to happen almost surely (sometimes abbreviated as a.s.) if it happens with probability 1 (or Lebesgue measure 1). In other words, the set of possible exceptions may be non-empty, but it has probability 0. ...
converge to an eigenvector corresponding to the eigenvalue of greatest magnitude, provided that has a nonzero component of this eigenvector in the eigenvector basis (and also provided that there is only one eigenvalue of greatest magnitude). This simple algorithm is useful in some practical applications; for example,
Google Google LLC () is an American multinational technology company focusing on search engine technology, online advertising, cloud computing, computer software, quantum computing, e-commerce, artificial intelligence, and consumer electronic ...
uses it to calculate the page rank of documents in their search engine. Also, the power method is the starting point for many more sophisticated algorithms. For instance, by keeping not just the last vector in the sequence, but instead looking at the span of ''all'' the vectors in the sequence, one can get a better (faster converging) approximation for the eigenvector, and this idea is the basis of
Arnoldi iteration In numerical linear algebra, the Arnoldi iteration is an eigenvalue algorithm and an important example of an iterative method. Arnoldi finds an approximation to the eigenvalues and eigenvectors of general (possibly non-Hermitian) matrices by cons ...
. Alternatively, the important
QR algorithm In numerical linear algebra, the QR algorithm or QR iteration is an eigenvalue algorithm: that is, a procedure to calculate the eigenvalues and eigenvectors of a matrix. The QR algorithm was developed in the late 1950s by John G. F. Francis and by ...
is also based on a subtle transformation of a power method.


Numerical computation of eigenvectors

Once the eigenvalues are computed, the eigenvectors could be calculated by solving the equation :\left(\mathbf - \lambda_i \mathbf\right)\mathbf_ = \mathbf using
Gaussian elimination In mathematics, Gaussian elimination, also known as row reduction, is an algorithm for solving systems of linear equations. It consists of a sequence of operations performed on the corresponding matrix of coefficients. This method can also be used ...
or any other method for solving
matrix equation In mathematics, a matrix (plural matrices) is a rectangular array or table of numbers, symbols, or expressions, arranged in rows and columns, which is used to represent a mathematical object or a property of such an object. For example, \beg ...
s. However, in practical large-scale eigenvalue methods, the eigenvectors are usually computed in other ways, as a byproduct of the eigenvalue computation. In
power iteration In mathematics, power iteration (also known as the power method) is an eigenvalue algorithm: given a diagonalizable matrix A, the algorithm will produce a number \lambda, which is the greatest (in absolute value) eigenvalue of A, and a nonzero vect ...
, for example, the eigenvector is actually computed before the eigenvalue (which is typically computed by the
Rayleigh quotient In mathematics, the Rayleigh quotient () for a given complex Hermitian matrix ''M'' and nonzero vector ''x'' is defined as: R(M,x) = . For real matrices and vectors, the condition of being Hermitian reduces to that of being symmetric, and the con ...
of the eigenvector). In the QR algorithm for a
Hermitian matrix In mathematics, a Hermitian matrix (or self-adjoint matrix) is a complex square matrix that is equal to its own conjugate transpose—that is, the element in the -th row and -th column is equal to the complex conjugate of the element in the ...
(or any normal matrix), the orthonormal eigenvectors are obtained as a product of the matrices from the steps in the algorithm. (For more general matrices, the QR algorithm yields the
Schur decomposition In the mathematical discipline of linear algebra, the Schur decomposition or Schur triangulation, named after Issai Schur, is a matrix decomposition. It allows one to write an arbitrary complex square matrix as unitarily equivalent to an upper tr ...
first, from which the eigenvectors can be obtained by a backsubstitution procedure.) For Hermitian matrices, the Divide-and-conquer eigenvalue algorithm is more efficient than the QR algorithm if both eigenvectors and eigenvalues are desired.


Additional topics


Generalized eigenspaces

Recall that the ''geometric'' multiplicity of an eigenvalue can be described as the dimension of the associated eigenspace, the
nullspace In mathematics, the kernel of a linear map, also known as the null space or nullspace, is the linear subspace of the domain of the map which is mapped to the zero vector. That is, given a linear map between two vector spaces and , the kerne ...
of . The algebraic multiplicity can also be thought of as a dimension: it is the dimension of the associated generalized eigenspace (1st sense), which is the nullspace of the matrix for ''any sufficiently large ''. That is, it is the space of ''
generalized eigenvector In linear algebra, a generalized eigenvector of an n\times n matrix A is a vector which satisfies certain criteria which are more relaxed than those for an (ordinary) eigenvector. Let V be an n-dimensional vector space; let \phi be a linear map ...
s'' (first sense), where a generalized eigenvector is any vector which ''eventually'' becomes 0 if is applied to it enough times successively. Any eigenvector is a generalized eigenvector, and so each eigenspace is contained in the associated generalized eigenspace. This provides an easy proof that the geometric multiplicity is always less than or equal to the algebraic multiplicity. This usage should not be confused with the ''generalized eigenvalue problem'' described below.


Conjugate eigenvector

A conjugate eigenvector or coneigenvector is a vector sent after transformation to a scalar multiple of its conjugate, where the scalar is called the conjugate eigenvalue or coneigenvalue of the linear transformation. The coneigenvectors and coneigenvalues represent essentially the same information and meaning as the regular eigenvectors and eigenvalues, but arise when an alternative coordinate system is used. The corresponding equation is : \mathbf\mathbf = \lambda \mathbf^*. For example, in coherent electromagnetic scattering theory, the linear transformation represents the action performed by the scattering object, and the eigenvectors represent polarization states of the electromagnetic wave. In
optics Optics is the branch of physics that studies the behaviour and properties of light, including its interactions with matter and the construction of instruments that use or detect it. Optics usually describes the behaviour of visible, ultravio ...
, the coordinate system is defined from the wave's viewpoint, known as the Forward Scattering Alignment (FSA), and gives rise to a regular eigenvalue equation, whereas in
radar Radar is a detection system that uses radio waves to determine the distance (''ranging''), angle, and radial velocity of objects relative to the site. It can be used to detect aircraft, ships, spacecraft, guided missiles, motor vehicles, ...
, the coordinate system is defined from the radar's viewpoint, known as the Back Scattering Alignment (BSA), and gives rise to a coneigenvalue equation.


Generalized eigenvalue problem

A generalized eigenvalue problem (second sense) is the problem of finding a (nonzero) vector that obeys : \mathbf\mathbf = \lambda \mathbf \mathbf where and are matrices. If obeys this equation, with some , then we call the ''generalized eigenvector'' of and (in the second sense), and is called the ''generalized eigenvalue'' of and (in the second sense) which corresponds to the generalized eigenvector . The possible values of must obey the following equation :\det(\mathbf - \lambda \mathbf)=0. If linearly independent vectors can be found, such that for every , , then we define the matrices and such that :P = \begin , & & , \\ \mathbf_1 & \cdots & \mathbf_n \\ , & & , \end \equiv \begin (\mathbf_1)_1 & \cdots & (\mathbf_n)_1 \\ \vdots & & \vdots \\ (\mathbf_1)_n & \cdots & (\mathbf_n)_n \end :(D)_ = \begin \lambda_i, & \texti = j\\ 0, & \text \end Then the following equality holds :\mathbf = \mathbf\mathbf\mathbf\mathbf^ And the proof is : \mathbf\mathbf= \mathbf \begin , & & , \\ \mathbf_1 & \cdots & \mathbf_n \\ , & & , \end = \begin , & & , \\ A\mathbf_1 & \cdots & A\mathbf_n \\ , & & , \end = \begin , & & , \\ \lambda_1B\mathbf_1 & \cdots & \lambda_nB\mathbf_n \\ , & & , \end = \begin , & & , \\ B\mathbf_1 & \cdots & B\mathbf_n \\ , & & , \end \mathbf = \mathbf\mathbf\mathbf And since is invertible, we multiply the equation from the right by its inverse, finishing the proof. The set of matrices of the form , where is a complex number, is called a ''pencil''; the term '' matrix pencil'' can also refer to the pair of matrices. If is invertible, then the original problem can be written in the form : \mathbf^\mathbf\mathbf = \lambda \mathbf which is a standard eigenvalue problem. However, in most situations it is preferable not to perform the inversion, but rather to solve the generalized eigenvalue problem as stated originally. This is especially important if and are
Hermitian matrices In mathematics, a Hermitian matrix (or self-adjoint matrix) is a complex square matrix that is equal to its own conjugate transpose—that is, the element in the -th row and -th column is equal to the complex conjugate of the element in the - ...
, since in this case is not generally Hermitian and important properties of the solution are no longer apparent. If and are both symmetric or Hermitian, and is also a
positive-definite matrix In mathematics, a symmetric matrix M with real entries is positive-definite if the real number z^\textsfMz is positive for every nonzero real column vector z, where z^\textsf is the transpose of More generally, a Hermitian matrix (that is, a ...
, the eigenvalues are real and eigenvectors and with distinct eigenvalues are -orthogonal (). In this case, eigenvectors can be chosen so that the matrix defined above satisfies :\mathbf^* \mathbf B \mathbf = \mathbf or \mathbf\mathbf^*\mathbf B = \mathbf, and there exists a
basis Basis may refer to: Finance and accounting *Adjusted basis, the net cost of an asset after adjusting for various tax-related items *Basis point, 0.01%, often used in the context of interest rates *Basis trading, a trading strategy consisting of ...
of generalized eigenvectors (it is not a defective problem). This case is sometimes called a ''Hermitian definite pencil'' or ''definite pencil''.


See also

* Eigenvalue perturbation *
Frobenius covariant In matrix theory, the Frobenius covariants of a square matrix are special polynomials of it, namely projection matrices ''A'i'' associated with the eigenvalues and eigenvectors of .Roger A. Horn and Charles R. Johnson (1991), ''Topics in M ...
*
Householder transformation In linear algebra, a Householder transformation (also known as a Householder reflection or elementary reflector) is a linear transformation that describes a reflection about a plane or hyperplane containing the origin. The Householder transformat ...
*
Jordan normal form In linear algebra, a Jordan normal form, also known as a Jordan canonical form (JCF), is an upper triangular matrix of a particular form called a Jordan matrix representing a linear operator on a finite-dimensional vector space with respect to ...
*
List of matrices This article lists some important classes of matrices used in mathematics, science and engineering. A matrix (plural matrices, or less commonly matrixes) is a rectangular array of numbers called ''entries''. Matrices have a long history of both st ...
*
Matrix decomposition In the mathematical discipline of linear algebra, a matrix decomposition or matrix factorization is a factorization of a matrix into a product of matrices. There are many different matrix decompositions; each finds use among a particular class of ...
*
Singular value decomposition In linear algebra, the singular value decomposition (SVD) is a factorization of a real or complex matrix. It generalizes the eigendecomposition of a square normal matrix with an orthonormal eigenbasis to any \ m \times n\ matrix. It is rela ...
*
Sylvester's formula In matrix theory, Sylvester's formula or Sylvester's matrix theorem (named after J. J. Sylvester) or Lagrange−Sylvester interpolation expresses an analytic function of a matrix as a polynomial in , in terms of the eigenvalues and eigenvectors ...


Notes


References

* * * * * * * {{cite book, last=Strang , first=G. , year=1998, title=Introduction to Linear Algebra, edition=3rd , publisher= Wellesley-Cambridge Press, isbn =978-0-9614088-5-5


External links


Interactive program & tutorial of Spectral Decomposition
Matrix theory Matrix decompositions