HOME

TheInfoList



OR:

In linear algebra, two ''n''-by-''n'' matrices and are called similar if there exists an invertible ''n''-by-''n'' matrix such that B = P^ A P . Similar matrices represent the same linear map under two (possibly) different bases, with being the
change of basis In mathematics, an ordered basis of a vector space of finite dimension allows representing uniquely any element of the vector space by a coordinate vector, which is a sequence of scalars called coordinates. If two different bases are consi ...
matrix. A transformation is called a similarity transformation or conjugation of the matrix . In the general linear group, similarity is therefore the same as
conjugacy In mathematics, especially group theory, two elements a and b of a group are conjugate if there is an element g in the group such that b = gag^. This is an equivalence relation whose equivalence classes are called conjugacy classes. In other wo ...
, and similar matrices are also called conjugate; however, in a given subgroup of the general linear group, the notion of conjugacy may be more restrictive than similarity, since it requires that be chosen to lie in .


Motivating example

When defining a linear transformation, it can be the case that a change of basis can result in a simpler form of the same transformation. For example, the matrix representing a rotation in when the
axis of rotation Rotation around a fixed axis is a special case of rotational motion. The fixed-axis hypothesis excludes the possibility of an axis changing its orientation and cannot describe such phenomena as wobbling or precession. According to Euler's rota ...
is not aligned with the coordinate axis can be complicated to compute. If the axis of rotation were aligned with the positive -axis, then it would simply be S = \begin \cos\theta & -\sin\theta & 0 \\ \sin\theta & \cos\theta & 0 \\ 0 & 0 & 1 \end, where \theta is the angle of rotation. In the new coordinate system, the transformation would be written as y' = Sx', where and are respectively the original and transformed vectors in a new basis containing a vector parallel to the axis of rotation. In the original basis, the transform would be written as y = Tx, where vectors and and the unknown transform matrix are in the original basis. To write in terms of the simpler matrix, we use the change-of-basis matrix that transforms and as x' = Px and y' = Py: \begin & &y' &= Sx' \\ &\Rightarrow &Py &= SPx \\ &\Rightarrow & y &= \left(P^SP\right)x = Tx \end Thus, the matrix in the original basis, T, is given by T = P^SP. The transform in the original basis is found to be the product of three easy-to-derive matrices. In effect, the similarity transform operates in three steps: change to a new basis (), perform the simple transformation (), and change back to the old basis ().


Properties

Similarity is an equivalence relation on the space of square matrices. Because matrices are similar if and only if they represent the same linear operator with respect to (possibly) different bases, similar matrices share all properties of their shared underlying operator: *
Rank Rank is the relative position, value, worth, complexity, power, importance, authority, level, etc. of a person or object within a ranking, such as: Level or position in a hierarchical organization * Academic rank * Diplomatic rank * Hierarchy * H ...
*
Characteristic polynomial In linear algebra, the characteristic polynomial of a square matrix is a polynomial which is invariant under matrix similarity and has the eigenvalues as roots. It has the determinant and the trace of the matrix among its coefficients. The char ...
, and attributes that can be derived from it: ** Determinant ** Trace ** Eigenvalues, and their algebraic multiplicities * Geometric multiplicities of eigenvalues (but not the eigenspaces, which are transformed according to the base change matrix ''P'' used). * Minimal polynomial *
Frobenius normal form In linear algebra, the Frobenius normal form or rational canonical form of a square matrix ''A'' with entries in a field ''F'' is a canonical form for matrices obtained by conjugation by invertible matrices over ''F''. The form reflects a minimal d ...
* Jordan normal form, up to a permutation of the Jordan blocks * Index of nilpotence * Elementary divisors, which form a complete set of invariants for similarity of matrices over a
principal ideal domain In mathematics, a principal ideal domain, or PID, is an integral domain in which every ideal is principal, i.e., can be generated by a single element. More generally, a principal ideal ring is a nonzero commutative ring whose ideals are princi ...
Because of this, for a given matrix ''A'', one is interested in finding a simple "normal form" ''B'' which is similar to ''A''—the study of ''A'' then reduces to the study of the simpler matrix ''B''. For example, ''A'' is called
diagonalizable In linear algebra, a square matrix A is called diagonalizable or non-defective if it is similar to a diagonal matrix, i.e., if there exists an invertible matrix P and a diagonal matrix D such that or equivalently (Such D are not unique.) F ...
if it is similar to a diagonal matrix. Not all matrices are diagonalizable, but at least over the complex numbers (or any
algebraically closed field In mathematics, a field is algebraically closed if every non-constant polynomial in (the univariate polynomial ring with coefficients in ) has a root in . Examples As an example, the field of real numbers is not algebraically closed, because ...
), every matrix is similar to a matrix in Jordan form. Neither of these forms is unique (diagonal entries or Jordan blocks may be permuted) so they are not really normal forms; moreover their determination depends on being able to factor the minimal or characteristic polynomial of ''A'' (equivalently to find its eigenvalues). The rational canonical form does not have these drawbacks: it exists over any field, is truly unique, and it can be computed using only arithmetic operations in the field; ''A'' and ''B'' are similar if and only if they have the same rational canonical form. The rational canonical form is determined by the elementary divisors of ''A''; these can be immediately read off from a matrix in Jordan form, but they can also be determined directly for any matrix by computing the Smith normal form, over the ring of polynomials, of the matrix (with polynomial entries) (the same one whose determinant defines the characteristic polynomial). Note that this Smith normal form is not a normal form of ''A'' itself; moreover it is not similar to either, but obtained from the latter by left and right multiplications by different invertible matrices (with polynomial entries). Similarity of matrices does not depend on the base field: if ''L'' is a field containing ''K'' as a subfield, and ''A'' and ''B'' are two matrices over ''K'', then ''A'' and ''B'' are similar as matrices over ''K'' if and only if they are similar as matrices over ''L''. This is so because the rational canonical form over ''K'' is also the rational canonical form over ''L''. This means that one may use Jordan forms that only exist over a larger field to determine whether the given matrices are similar. In the definition of similarity, if the matrix ''P'' can be chosen to be a permutation matrix then ''A'' and ''B'' are permutation-similar; if ''P'' can be chosen to be a
unitary matrix In linear algebra, a complex square matrix is unitary if its conjugate transpose is also its inverse, that is, if U^* U = UU^* = UU^ = I, where is the identity matrix. In physics, especially in quantum mechanics, the conjugate transpose ...
then ''A'' and ''B'' are unitarily equivalent. The
spectral theorem In mathematics, particularly linear algebra and functional analysis, a spectral theorem is a result about when a linear operator or matrix can be diagonalized (that is, represented as a diagonal matrix in some basis). This is extremely useful be ...
says that every
normal matrix In mathematics, a complex square matrix is normal if it commutes with its conjugate transpose : The concept of normal matrices can be extended to normal operators on infinite dimensional normed spaces and to normal elements in C*-algebras. A ...
is unitarily equivalent to some diagonal matrix.
Specht's theorem In mathematics, Specht's theorem gives a necessary and sufficient condition for two complex matrices to be unitarily equivalent. It is named after Wilhelm Specht, who proved the theorem in 1940. Two matrices ''A'' and ''B'' with complex number ...
states that two matrices are unitarily equivalent if and only if they satisfy certain trace equalities.


See also

* Canonical forms * Matrix congruence * Matrix equivalence


Notes


References

* * * Horn and Johnson, ''Matrix Analysis,'' Cambridge University Press, 1985. {{isbn, 0-521-38632-2. (Similarity is discussed many places, starting at page 44.) Matrices Equivalence (mathematics)