HOME

TheInfoList



OR:

In
linear algebra Linear algebra is the branch of mathematics concerning linear equations such as :a_1x_1+\cdots +a_nx_n=b, linear maps such as :(x_1, \ldots, x_n) \mapsto a_1x_1+\cdots +a_nx_n, and their representations in vector spaces and through matrix (mathemat ...
, the characteristic polynomial of a square matrix is a
polynomial In mathematics, a polynomial is a Expression (mathematics), mathematical expression consisting of indeterminate (variable), indeterminates (also called variable (mathematics), variables) and coefficients, that involves only the operations of addit ...
which is invariant under matrix similarity and has the eigenvalues as
roots A root is the part of a plant, generally underground, that anchors the plant body, and absorbs and stores water and nutrients. Root or roots may also refer to: Art, entertainment, and media * ''The Root'' (magazine), an online magazine focusin ...
. It has the
determinant In mathematics, the determinant is a Scalar (mathematics), scalar-valued function (mathematics), function of the entries of a square matrix. The determinant of a matrix is commonly denoted , , or . Its value characterizes some properties of the ...
and the trace of the matrix among its coefficients. The characteristic polynomial of an endomorphism of a finite-dimensional
vector space In mathematics and physics, a vector space (also called a linear space) is a set (mathematics), set whose elements, often called vector (mathematics and physics), ''vectors'', can be added together and multiplied ("scaled") by numbers called sc ...
is the characteristic polynomial of the matrix of that endomorphism over any basis (that is, the characteristic polynomial does not depend on the choice of a basis). The characteristic equation, also known as the determinantal equation, is the equation obtained by equating the characteristic polynomial to zero. In spectral graph theory, the characteristic polynomial of a graph is the characteristic polynomial of its adjacency matrix.


Motivation

In
linear algebra Linear algebra is the branch of mathematics concerning linear equations such as :a_1x_1+\cdots +a_nx_n=b, linear maps such as :(x_1, \ldots, x_n) \mapsto a_1x_1+\cdots +a_nx_n, and their representations in vector spaces and through matrix (mathemat ...
,
eigenvalues and eigenvectors In linear algebra, an eigenvector ( ) or characteristic vector is a vector that has its direction unchanged (or reversed) by a given linear transformation. More precisely, an eigenvector \mathbf v of a linear transformation T is scaled by a ...
play a fundamental role, since, given a
linear transformation In mathematics, and more specifically in linear algebra, a linear map (also called a linear mapping, linear transformation, vector space homomorphism, or in some contexts linear function) is a mapping V \to W between two vector spaces that pr ...
, an eigenvector is a vector whose direction is not changed by the transformation, and the corresponding eigenvalue is the measure of the resulting change of magnitude of the vector. More precisely, suppose the transformation is represented by a square matrix A. Then an eigenvector \mathbf and the corresponding eigenvalue \lambda must satisfy the equation A \mathbf = \lambda \mathbf, or, equivalently (since \lambda \mathbf = \lambda I \mathbf), (\lambda I - A) \mathbf =\mathbf 0 where I is the
identity matrix In linear algebra, the identity matrix of size n is the n\times n square matrix with ones on the main diagonal and zeros elsewhere. It has unique properties, for example when the identity matrix represents a geometric transformation, the obje ...
, and \mathbf\ne \mathbf (although the zero vector satisfies this equation for every \lambda, it is not considered an eigenvector). It follows that the matrix (\lambda I - A) must be singular, and its determinant \det(\lambda I - A) = 0 must be zero. In other words, the eigenvalues of are the
roots A root is the part of a plant, generally underground, that anchors the plant body, and absorbs and stores water and nutrients. Root or roots may also refer to: Art, entertainment, and media * ''The Root'' (magazine), an online magazine focusin ...
of \det(xI - A), which is a
monic polynomial In algebra, a monic polynomial is a non-zero univariate polynomial (that is, a polynomial in a single variable) in which the leading coefficient (the nonzero coefficient of highest degree) is equal to 1. That is to say, a monic polynomial is one ...
in of degree if is a matrix. This polynomial is the ''characteristic polynomial'' of .


Formal definition

Consider an n \times n matrix A. The characteristic polynomial of A, denoted by p_A(t), is the polynomial defined by p_A(t) = \det (t I - A) where I denotes the n \times n
identity matrix In linear algebra, the identity matrix of size n is the n\times n square matrix with ones on the main diagonal and zeros elsewhere. It has unique properties, for example when the identity matrix represents a geometric transformation, the obje ...
. Some authors define the characteristic polynomial to be \det(A - t I). That polynomial differs from the one defined here by a sign (-1)^n, so it makes no difference for properties like having as roots the eigenvalues of A; however the definition above always gives a
monic polynomial In algebra, a monic polynomial is a non-zero univariate polynomial (that is, a polynomial in a single variable) in which the leading coefficient (the nonzero coefficient of highest degree) is equal to 1. That is to say, a monic polynomial is one ...
, whereas the alternative definition is monic only when n is even.


Examples

To compute the characteristic polynomial of the matrix A = \begin 2 & 1\\ -1& 0 \end. the
determinant In mathematics, the determinant is a Scalar (mathematics), scalar-valued function (mathematics), function of the entries of a square matrix. The determinant of a matrix is commonly denoted , , or . Its value characterizes some properties of the ...
of the following is computed: t I-A = \begin t-2&-1\\ 1&t-0 \end and found to be (t-2)t - 1(-1) = t^2-2t+1 \,\!, the characteristic polynomial of A. Another example uses hyperbolic functions of a hyperbolic angle φ. For the matrix take A = \begin \cosh(\varphi) & \sinh(\varphi)\\ \sinh(\varphi)& \cosh(\varphi) \end. Its characteristic polynomial is \det (tI - A) = (t - \cosh(\varphi))^2 - \sinh^2(\varphi) = t^2 - 2 t \ \cosh(\varphi) + 1 = (t - e^\varphi) (t - e^).


Properties

The characteristic polynomial p_A(t) of a n \times n matrix is monic (its leading coefficient is 1) and its degree is n. The most important fact about the characteristic polynomial was already mentioned in the motivational paragraph: the eigenvalues of A are precisely the
root In vascular plants, the roots are the plant organ, organs of a plant that are modified to provide anchorage for the plant and take in water and nutrients into the plant body, which allows plants to grow taller and faster. They are most often bel ...
s of p_A(t) (this also holds for the minimal polynomial of A, but its degree may be less than n). All coefficients of the characteristic polynomial are
polynomial expression In mathematics, especially in the field of algebra, a polynomial ring or polynomial algebra is a ring (mathematics), ring formed from the set (mathematics), set of polynomials in one or more indeterminate (variable), indeterminates (traditionally ...
s in the entries of the matrix. In particular its constant coefficient of t^0 is \det(-A) = (-1)^n \det(A), the coefficient of t^n is one, and the coefficient of t^ is , where is the trace of A. (The signs given here correspond to the formal definition given in the previous section; for the alternative definition these would instead be \det(A) and respectively.) For a 2 \times 2 matrix A, the characteristic polynomial is thus given by t^2 - \operatorname(A) t + \det(A). Using the language of
exterior algebra In mathematics, the exterior algebra or Grassmann algebra of a vector space V is an associative algebra that contains V, which has a product, called exterior product or wedge product and denoted with \wedge, such that v\wedge v=0 for every vector ...
, the characteristic polynomial of an n \times n matrix A may be expressed as p_A (t) = \sum_^n t^ (-1)^k \operatorname\left(\textstyle\bigwedge^k A\right) where \operatorname\left(\bigwedge^k A\right) is the trace of the kth exterior power of A, which has dimension \binom . This trace may be computed as the sum of all principal minors of A of size k. The recursive Faddeev–LeVerrier algorithm computes these coefficients more efficiently . When the characteristic of the field of the coefficients is 0, each such trace may alternatively be computed as a single determinant, that of the k \times k matrix, \operatorname\left(\textstyle\bigwedge^k A\right) = \frac \begin \operatornameA & k-1 &0&\cdots &0 \\ \operatornameA^2 &\operatornameA& k-2 &\cdots &0 \\ \vdots & \vdots & & \ddots & \vdots \\ \operatornameA^ &\operatornameA^& & \cdots & 1 \\ \operatornameA^k &\operatornameA^& & \cdots & \operatornameA \end ~. The Cayley–Hamilton theorem states that replacing t by A in the characteristic polynomial (interpreting the resulting powers as matrix powers, and the constant term c as c times the identity matrix) yields the zero matrix. Informally speaking, every matrix satisfies its own characteristic equation. This statement is equivalent to saying that the minimal polynomial of A divides the characteristic polynomial of A. Two
similar matrices In linear algebra, two ''n''-by-''n'' matrices and are called similar if there exists an invertible ''n''-by-''n'' matrix such that B = P^ A P . Similar matrices represent the same linear map under two possibly different bases, with being ...
have the same characteristic polynomial. The converse however is not true in general: two matrices with the same characteristic polynomial need not be similar. The matrix A and its
transpose In linear algebra, the transpose of a Matrix (mathematics), matrix is an operator which flips a matrix over its diagonal; that is, it switches the row and column indices of the matrix by producing another matrix, often denoted by (among other ...
have the same characteristic polynomial. A is similar to a triangular matrix
if and only if In logic and related fields such as mathematics and philosophy, "if and only if" (often shortened as "iff") is paraphrased by the biconditional, a logical connective between statements. The biconditional is true in two cases, where either bo ...
its characteristic polynomial can be completely factored into linear factors over K (the same is true with the minimal polynomial instead of the characteristic polynomial). In this case A is similar to a matrix in Jordan normal form.


Characteristic polynomial of a product of two matrices

If A and B are two square n \times n matrices then characteristic polynomials of AB and BA coincide: p_(t)=p_(t).\, When A is non-singular this result follows from the fact that AB and BA are similar: BA = A^ (AB) A. For the case where both A and B are singular, the desired identity is an equality between polynomials in t and the coefficients of the matrices. Thus, to prove this equality, it suffices to prove that it is verified on a non-empty open subset (for the usual
topology Topology (from the Greek language, Greek words , and ) is the branch of mathematics concerned with the properties of a Mathematical object, geometric object that are preserved under Continuous function, continuous Deformation theory, deformat ...
, or, more generally, for the Zariski topology) of the space of all the coefficients. As the non-singular matrices form such an open subset of the space of all matrices, this proves the result. More generally, if A is a matrix of order m \times n and B is a matrix of order n \times m, then AB is m \times m and BA is n \times n matrix, and one has p_(t) = t^ p_(t).\, To prove this, one may suppose n > m, by exchanging, if needed, A and B. Then, by bordering A on the bottom by n - m rows of zeros, and B on the right, by, n - m columns of zeros, one gets two n \times n matrices A^ and B^ such that B^A^ = BA and A^B^ is equal to AB bordered by n - m rows and columns of zeros. The result follows from the case of square matrices, by comparing the characteristic polynomials of A^B^ and AB.


Characteristic polynomial of ''A''''k''

If \lambda is an eigenvalue of a square matrix A with eigenvector \mathbf, then \lambda^k is an eigenvalue of A^k because A^k \textbf = A^ A \textbf = \lambda A^ \textbf = \dots = \lambda^k \textbf. The multiplicities can be shown to agree as well, and this generalizes to any polynomial in place of x^k: That is, the algebraic multiplicity of \lambda in f(A) equals the sum of algebraic multiplicities of \lambda' in A over \lambda' such that f(\lambda') = \lambda. In particular, \operatorname(f(A)) = \textstyle\sum_^n f(\lambda_i) and \operatorname(f(A)) = \textstyle\prod_^n f(\lambda_i). Here a polynomial f(t) = t^3+1, for example, is evaluated on a matrix A simply as f(A) = A^3+I. The theorem applies to matrices and polynomials over any field or
commutative ring In mathematics, a commutative ring is a Ring (mathematics), ring in which the multiplication operation is commutative. The study of commutative rings is called commutative algebra. Complementarily, noncommutative algebra is the study of ring prope ...
. However, the assumption that p_A(t) has a factorization into linear factors is not always true, unless the matrix is over an
algebraically closed field In mathematics, a field is algebraically closed if every non-constant polynomial in (the univariate polynomial ring with coefficients in ) has a root in . In other words, a field is algebraically closed if the fundamental theorem of algebra ...
such as the complex numbers.


Secular function and secular equation


Secular function

The term secular function has been used for what is now called ''characteristic polynomial'' (in some literature the term secular function is still used). The term comes from the fact that the characteristic polynomial was used to calculate secular perturbations (on a time scale of a century, that is, slow compared to annual motion) of planetary orbits, according to
Lagrange Joseph-Louis Lagrange (born Giuseppe Luigi Lagrangialinear algebra Linear algebra is the branch of mathematics concerning linear equations such as :a_1x_1+\cdots +a_nx_n=b, linear maps such as :(x_1, \ldots, x_n) \mapsto a_1x_1+\cdots +a_nx_n, and their representations in vector spaces and through matrix (mathemat ...
it is sometimes used in place of characteristic equation. * In
astronomy Astronomy is a natural science that studies celestial objects and the phenomena that occur in the cosmos. It uses mathematics, physics, and chemistry in order to explain their origin and their overall evolution. Objects of interest includ ...
it is the algebraic or numerical expression of the magnitude of the inequalities in a planet's motion that remain after the inequalities of a short period have been allowed for. * In molecular orbital calculations relating to the energy of the electron and its wave function it is also used instead of the characteristic equation.


For general associative algebras

The above definition of the characteristic polynomial of a matrix A \in M_n(F) with entries in a field F generalizes without any changes to the case when F is just a
commutative ring In mathematics, a commutative ring is a Ring (mathematics), ring in which the multiplication operation is commutative. The study of commutative rings is called commutative algebra. Complementarily, noncommutative algebra is the study of ring prope ...
. defines the characteristic polynomial for elements of an arbitrary finite-dimensional (
associative In mathematics, the associative property is a property of some binary operations that rearranging the parentheses in an expression will not change the result. In propositional logic, associativity is a valid rule of replacement for express ...
, but not necessarily commutative) algebra over a field F and proves the standard properties of the characteristic polynomial in this generality.


See also

* Characteristic equation (disambiguation) * Invariants of tensors * Companion matrix * Faddeev–LeVerrier algorithm * Cayley–Hamilton theorem * Samuelson–Berkowitz algorithm


References

* T.S. Blyth & E.F. Robertson (1998) ''Basic Linear Algebra'', p 149, Springer . * John B. Fraleigh & Raymond A. Beauregard (1990) ''Linear Algebra'' 2nd edition, p 246,
Addison-Wesley Addison–Wesley is an American publisher of textbooks and computer literature. It is an imprint of Pearson plc, a global publishing and education company. In addition to publishing books, Addison–Wesley also distributes its technical titles ...
. * * Werner Greub (1974) ''Linear Algebra'' 4th edition, pp 120–5, Springer, . * Paul C. Shields (1980) ''Elementary Linear Algebra'' 3rd edition, p 274, Worth Publishers . *
Gilbert Strang William Gilbert Strang (born November 27, 1934) is an American mathematician known for his contributions to Finite elements, finite element theory, the calculus of variations, wavelet analysis and linear algebra. He has made many contributions ...
(1988) ''Linear Algebra and Its Applications'' 3rd edition, p 246, Brooks/Cole {{ISBN, 0-15-551005-3 . Polynomials Linear algebra Tensors