HOME

TheInfoList



OR:

In
linear algebra Linear algebra is the branch of mathematics concerning linear equations such as :a_1x_1+\cdots +a_nx_n=b, linear maps such as :(x_1, \ldots, x_n) \mapsto a_1x_1+\cdots +a_nx_n, and their representations in vector spaces and through matrix (mathemat ...
, the
Frobenius Frobenius is a surname. Notable people with the surname include: * Ferdinand Georg Frobenius (1849–1917), mathematician ** Frobenius algebra ** Frobenius endomorphism ** Frobenius inner product ** Frobenius norm ** Frobenius method ** Frobenius g ...
companion matrix of the
monic polynomial In algebra, a monic polynomial is a non-zero univariate polynomial (that is, a polynomial in a single variable) in which the leading coefficient (the nonzero coefficient of highest degree) is equal to 1. That is to say, a monic polynomial is one ...
p(x)=c_0 + c_1 x + \cdots + c_x^ + x^n is the square matrix defined as C(p)=\begin 0 & 0 & \dots & 0 & -c_0 \\ 1 & 0 & \dots & 0 & -c_1 \\ 0 & 1 & \dots & 0 & -c_2 \\ \vdots & \vdots & \ddots & \vdots & \vdots \\ 0 & 0 & \dots & 1 & -c_ \end. Some authors use the
transpose In linear algebra, the transpose of a Matrix (mathematics), matrix is an operator which flips a matrix over its diagonal; that is, it switches the row and column indices of the matrix by producing another matrix, often denoted by (among other ...
of this matrix, C(p)^T , which is more convenient for some purposes such as linear
recurrence relation In mathematics, a recurrence relation is an equation according to which the nth term of a sequence of numbers is equal to some combination of the previous terms. Often, only k previous terms of the sequence appear in the equation, for a parameter ...
s ( see below). C(p) is defined from the coefficients of p(x), while the
characteristic polynomial In linear algebra, the characteristic polynomial of a square matrix is a polynomial which is invariant under matrix similarity and has the eigenvalues as roots. It has the determinant and the trace of the matrix among its coefficients. The ...
as well as the minimal polynomial of C(p) are equal to p(x) . In this sense, the matrix C(p) and the polynomial p(x) are "companions".


Similarity to companion matrix

Any matrix with entries in a field has characteristic polynomial p(x) = \det(xI - A) , which in turn has companion matrix C(p) . These matrices are related as follows. The following statements are equivalent: * ''A'' is similar over ''F'' to C(p) , i.e. ''A'' can be conjugated to its companion matrix by matrices in GL''n''(''F''); * the characteristic polynomial p(x) coincides with the minimal polynomial of ''A'' , i.e. the minimal polynomial has degree ''n''; * the linear mapping A:F^n\to F^n makes F^n a cyclic F /math>-module, having a basis of the form \ ; or equivalently F^n \cong F (p(x)) as F /math>-modules. If the above hold, one says that ''A'' is ''non-derogatory''. Not every square matrix is similar to a companion matrix, but every square matrix is similar to a block diagonal matrix made of companion matrices. If we also demand that the polynomial of each diagonal block divides the next one, they are uniquely determined by ''A'', and this gives the rational canonical form of ''A''.


Diagonalizability

The roots of the characteristic polynomial p(x) are the
eigenvalue In linear algebra, an eigenvector ( ) or characteristic vector is a vector that has its direction unchanged (or reversed) by a given linear transformation. More precisely, an eigenvector \mathbf v of a linear transformation T is scaled by a ...
s of C(p) . If there are ''n'' distinct eigenvalues \lambda_1,\ldots,\lambda_n, then C(p) is diagonalizable as C(p) = V^\!D V, where ''D'' is the diagonal matrix and ''V'' is the Vandermonde matrix corresponding to the 's: D = \begin \lambda_1 & 0 & \!\!\!\cdots\!\!\! & 0\\ 0 & \lambda_2 & \!\!\!\cdots\!\!\! & 0\\ 0 & 0 & \!\!\!\cdots\!\!\! & \lambda_n \end, \qquad V = \begin 1 & \lambda_1 & \lambda_1^2 & \!\!\!\cdots\!\!\! & \lambda_1^\\ 1 & \lambda_2 & \lambda_2^2 & \!\!\!\cdots\!\!\! & \lambda_2^\\ 1em\vdots & \vdots & \vdots & \!\!\!\ddots\!\!\! &\vdots \\ 1 & \lambda_n & \lambda_n^2 & \!\!\!\cdots\!\!\! & \lambda_n^ \end. Indeed, a reasonably hard computation shows that the transpose C(p)^T has eigenvectors v_i=(1,\lambda_i,\ldots,\lambda_i ^) with C(p)^T\!(v_i) = \lambda_i v_i , which follows from p(\lambda_i)=c_0+c_1\lambda_i+\cdots+c_\lambda_i^+\lambda_i ^n = 0 . Thus, its diagonalizing change of basis matrix is V^T = _1^T \ldots v_n^T, meaning C(p)^T = V^T D\, (V^T) ^ , and taking the transpose of both sides gives C(p) = V^\!D V. We can read the eigenvectors of C(p) with C(p)(w_i) = \lambda_i w_i from the equation C(p) = V^\!D V: they are the column vectors of the inverse Vandermonde matrix V^ = _1^T \cdots w_n^T. This matrix is known explicitly, giving the eigenvectors w_i = (L_, \ldots, L_) , with coordinates equal to the coefficients of the Lagrange polynomials L_i(x) = L_ + L_x + \cdots + L_x^ = \prod_ \frac = \frac. Alternatively, the scaled eigenvectors \tilde w_i = p'\!(\lambda_i)\,w_i have simpler coefficients. If p(x) has multiple roots, then C(p) is not diagonalizable. Rather, the Jordan canonical form of C(p) contains one Jordan block for each distinct root; if the multiplicity of the root is ''m'', then the block is an ''m'' × ''m'' matrix with \lambda on the diagonal and 1 in the entries just above the diagonal. in this case, ''V'' becomes a confluent Vandermonde matrix.


Linear recursive sequences

A linear recursive sequence defined by a_ = - c_0 a_k - c_1 a_ \cdots - c_ a_ for k \geq 0 has the characteristic polynomial p(x)=c_0 + c_1 x + \cdots + c_x^ + x^n , whose transpose companion matrix C(p)^T generates the sequence: \begina_\\ a_\\ \vdots \\ a_\\ a_ \end = \begin 0 & 1 & 0 & \cdots & 0\\ 0 & 0 & 1 & \cdots & 0\\ \vdots & \vdots & \vdots & \ddots & \vdots \\ 0 & 0 & 0 & \cdots & 1\\ -c_0 & -c_1 & -c_2 & \cdots & -c_ \end \begina_k\\ a_\\ \vdots \\ a_\\ a_ \end . The vector v=(1,\lambda,\lambda^2,\ldots,\lambda^) is an eigenvector of this matrix, where the eigenvalue \lambda is a root of p(x). Setting the initial values of the sequence equal to this vector produces a geometric sequence a_k = \lambda^k which satisfies the recurrence. In the case of ''n'' distinct eigenvalues, an arbitrary solution a_k can be written as a linear combination of such geometric solutions, and the eigenvalues of largest complex norm give an asymptotic approximation.


From linear ODE to first-order linear ODE system

Similarly to the above case of linear recursions, consider a homogeneous linear ODE of order ''n'' for the scalar function y = y(t): y^ + c_y^ + \dots + c_y^ + c_0 y = 0 . This can be equivalently described as a coupled system of homogeneous linear ODE of order 1 for the vector function z(t) = (y(t),y'(t),\ldots,y^(t)): z' = C(p)^T z where C(p)^T is the transpose companion matrix for the characteristic polynomial p(x)=x^n + c_x^ + \cdots + c_1 x + c_0 . Here the coefficients c_i = c_i(t) may be also functions, not just constants. If C(p)^T is diagonalizable, then a diagonalizing change of basis will transform this into a decoupled system equivalent to one scalar homogeneous first-order linear ODE in each coordinate. An inhomogeneous equation y^ + c_y^ + \dots + c_y^ + c_0 y = f(t) is equivalent to the system: z' = C(p)^T z + F(t) with the inhomogeneity term F(t) = (0,\ldots,0,f(t)) . Again, a diagonalizing change of basis will transform this into a decoupled system of scalar inhomogeneous first-order linear ODEs.


Cyclic shift matrix

In the case of p(x)=x^n-1, when the eigenvalues are the complex
roots of unity In mathematics, a root of unity is any complex number that yields 1 when raised to some positive integer power . Roots of unity are used in many branches of mathematics, and are especially important in number theory, the theory of group char ...
, the companion matrix and its transpose both reduce to Sylvester's cyclic shift matrix, a circulant matrix.


Multiplication map on a simple field extension

Consider a polynomial p(x)=x^n + c_x^ + \cdots + c_1 x + c_0 with coefficients in a field F, and suppose p(x) is irreducible in the
polynomial ring In mathematics, especially in the field of algebra, a polynomial ring or polynomial algebra is a ring formed from the set of polynomials in one or more indeterminates (traditionally also called variables) with coefficients in another ring, ...
F /math>. Then adjoining a root \lambda of p(x) produces a
field extension In mathematics, particularly in algebra, a field extension is a pair of fields K \subseteq L, such that the operations of ''K'' are those of ''L'' restricted to ''K''. In this case, ''L'' is an extension field of ''K'' and ''K'' is a subfield of ...
K=F(\lambda) \cong F (p(x)), which is also a vector space over F with standard basis \ . Then the F-linear multiplication mapping has an ''n'' × ''n'' matrix _\lambda/math> with respect to the standard basis. Since m_\lambda(\lambda^i) = \lambda^ and m_\lambda(\lambda^) = \lambda^n = -c_0-\cdots-c_\lambda^, this is the companion matrix of p(x): _\lambda= C(p). Assuming this extension is separable (for example if F has characteristic zero or is a
finite field In mathematics, a finite field or Galois field (so-named in honor of Évariste Galois) is a field (mathematics), field that contains a finite number of Element (mathematics), elements. As with any field, a finite field is a Set (mathematics), s ...
), p(x) has distinct roots \lambda_1,\ldots,\lambda_n with \lambda_1 = \lambda, so that p(x)=(x-\lambda_1)\cdots (x-\lambda_n), and it has
splitting field In abstract algebra, a splitting field of a polynomial with coefficients in a field is the smallest field extension of that field over which the polynomial ''splits'', i.e., decomposes into linear factors. Definition A splitting field of a polyn ...
L = F(\lambda_1,\ldots,\lambda_n). Now m_\lambda is not diagonalizable over F; rather, we must extend it to an L-linear map on L^n \cong L\otimes_F K, a vector space over L with standard basis \ , containing vectors w = (\beta_1,\ldots,\beta_n) = \beta_1 1+\cdots+\beta_n \lambda^ . The extended mapping is defined by m_\lambda(\beta\otimes\alpha) = \beta\otimes(\lambda\alpha). The matrix _\lambda= C(p) is unchanged, but as above, it can be diagonalized by matrices with entries in L: _\lambdaC(p)= V^\! D V, for the diagonal matrix D=\operatorname(\lambda_1,\ldots,\lambda_n) and the Vandermonde matrix ''V'' corresponding to \lambda_1,\ldots,\lambda_n\in L . The explicit formula for the eigenvectors (the scaled column vectors of the inverse Vandermonde matrix V^) can be written as: \tilde w_i = \beta_ 1+\beta_ \lambda +\cdots + \beta_ \lambda^ = \prod_ (1\lambda - \lambda_j 1) where \beta_\in L are the coefficients of the scaled Lagrange polynomial \frac = \prod_ (x - \lambda_j) = \beta_ + \beta_ x + \cdots + \beta_ x^.


See also

*
Frobenius endomorphism In commutative algebra and field theory (mathematics), field theory, the Frobenius endomorphism (after Ferdinand Georg Frobenius) is a special endomorphism of commutative Ring (mathematics), rings with prime number, prime characteristic (algebra), ...
* Cayley–Hamilton theorem * Krylov subspace


Notes

{{DEFAULTSORT:Companion Matrix Matrices (mathematics) Matrix theory Matrix normal forms