HOME





Nilpotent Matrix
In linear algebra, a nilpotent matrix is a square matrix ''N'' such that :N^k = 0\, for some positive integer k. The smallest such k is called the index of N, sometimes the degree of N. More generally, a nilpotent transformation is a linear transformation L of a vector space such that L^k = 0 for some positive integer k (and thus, L^j = 0 for all j \geq k). Both of these concepts are special cases of a more general concept of nilpotent, nilpotence that applies to elements of ring (algebra), rings. Examples Example 1 The matrix : A = \begin 0 & 1 \\ 0 & 0 \end is nilpotent with index 2, since A^2 = 0. Example 2 More generally, any n-dimensional triangular matrix with zeros along the main diagonal is nilpotent, with index \le n . For example, the matrix : B=\begin 0 & 2 & 1 & 6\\ 0 & 0 & 1 & 2\\ 0 & 0 & 0 & 3\\ 0 & 0 & 0 & 0 \end is nilpotent, with : B^2=\begin 0 & 0 & 2 & 7\\ 0 & 0 & 0 & 3\\ 0 & 0 & 0 & 0\\ 0 & 0 & 0 & 0 \end ;\ B^3=\begin 0 & 0 & 0 & 6\\ 0 & 0 & 0 & ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


picture info

Linear Algebra
Linear algebra is the branch of mathematics concerning linear equations such as :a_1x_1+\cdots +a_nx_n=b, linear maps such as :(x_1, \ldots, x_n) \mapsto a_1x_1+\cdots +a_nx_n, and their representations in vector spaces and through matrix (mathematics), matrices. Linear algebra is central to almost all areas of mathematics. For instance, linear algebra is fundamental in modern presentations of geometry, including for defining basic objects such as line (geometry), lines, plane (geometry), planes and rotation (mathematics), rotations. Also, functional analysis, a branch of mathematical analysis, may be viewed as the application of linear algebra to Space of functions, function spaces. Linear algebra is also used in most sciences and fields of engineering because it allows mathematical model, modeling many natural phenomena, and computing efficiently with such models. For nonlinear systems, which cannot be modeled with linear algebra, it is often used for dealing with first-order a ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


Newton's Identities
In mathematics, Newton's identities, also known as the Girard–Newton formulae, give relations between two types of symmetric polynomials, namely between power sums and elementary symmetric polynomials. Evaluated at the roots of a monic polynomial ''P'' in one variable, they allow expressing the sums of the ''k''-th powers of all roots of ''P'' (counted with their multiplicity) in terms of the coefficients of ''P'', without actually finding those roots. These identities were found by Isaac Newton around 1666, apparently in ignorance of earlier work (1629) by Albert Girard. They have applications in many areas of mathematics, including Galois theory, invariant theory, group theory, combinatorics, as well as further applications outside mathematics, including general relativity. Mathematical statement Formulation in terms of symmetric polynomials Let ''x''1, ..., ''x''''n'' be variables, denote for ''k'' ≥ 1 by ''p''''k''(''x''1, ..., ''x''''n'') the ''k''-th p ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


Flag (linear Algebra)
In mathematics, particularly in linear algebra, a flag is an increasing sequence of subspaces of a finite-dimensional vector space ''V''. Here "increasing" means each is a proper subspace of the next (see filtration): :\ = V_0 \sub V_1 \sub V_2 \sub \cdots \sub V_k = V. The term ''flag'' is motivated by a particular example resembling a flag: the zero point, a line, and a plane correspond to a nail, a staff, and a sheet of fabric. If we write that dim''V''''i'' = ''d''''i'' then we have :0 = d_0 < d_1 < d_2 < \cdots < d_k = n, where ''n'' is the of ''V'' (assumed to be finite). Hence, we must have ''k'' ≤ ''n''. A flag is called a complete flag if ''d''''i'' = ''i'' for all ''i'', otherwise it is called a partial flag. A partial flag can be obtained from a complete flag by deleting some of the subspace ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


Classification Theorem
In mathematics, a classification theorem answers the classification problem: "What are the objects of a given type, up to some equivalence?". It gives a non-redundant enumeration: each object is equivalent to exactly one class. A few issues related to classification are the following. *The equivalence problem is "given two objects, determine if they are equivalent". *A complete set of invariants, together with which invariants are realizable, solves the classification problem, and is often a step in solving it. (A combination of invariant values is realizable if there in fact exists an object whose invariants take on the specified set of values) *A (together with which invariants are realizable) solves both the classification problem and the equivalence problem. * A canonical form solves the classification problem, and is more data: it not only classifies every class, but provides a distinguished (canonical) element of each class. There exist many classification theorems in mat ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  




Jordan Canonical Form
\begin \lambda_1 1\hphantom\hphantom\\ \hphantom\lambda_1 1\hphantom\\ \hphantom\lambda_1\hphantom\\ \hphantom\lambda_2 1\hphantom\hphantom\\ \hphantom\hphantom\lambda_2\hphantom\\ \hphantom\lambda_3\hphantom\\ \hphantom\ddots\hphantom\\ \hphantom\lambda_n 1\hphantom\\ \hphantom\hphantom\lambda_n \end Example of a matrix in Jordan normal form. All matrix entries not shown are zero. The outlined squares are known as "Jordan blocks". Each Jordan block contains one number ''λi'' on its main diagonal, and 1s directly above the main diagonal. The ''λi''s are the eigenvalues of the matrix; they need not be distinct. In linear algebra, a Jordan normal form, also known as a Jordan canonical form, is an upper triangular matrix of a particular form called a Jordan matrix representing a linear operator on a finite-dimensional vector space with respect to some basis. Such a matrix has each non-zero off-diagonal entry equal to 1, immediately above the main diagonal (on the ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


picture info

Block Diagonal Matrix
In mathematics, a block matrix or a partitioned matrix is a matrix that is interpreted as having been broken into sections called blocks or submatrices. Intuitively, a matrix interpreted as a block matrix can be visualized as the original matrix with a collection of horizontal and vertical lines, which break it up, or partition it, into a collection of smaller matrices. For example, the 3x4 matrix presented below is divided by horizontal and vertical lines into four blocks: the top-left 2x3 block, the top-right 2x1 block, the bottom-left 1x3 block, and the bottom-right 1x1 block. : \left \begin a_ & a_ & a_ & b_ \\ a_ & a_ & a_ & b_ \\ \hline c_ & c_ & c_ & d \end \right Any matrix may be interpreted as a block matrix in one or more ways, with each interpretation defined by how its rows and columns are partitioned. This notion can be made more precise for an n by m matrix M by partitioning n into a collection \text, and then partitioning m into a collection \text. The original m ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


Matrix Similarity
In linear algebra, two ''n''-by-''n'' matrices and are called similar if there exists an invertible ''n''-by-''n'' matrix such that B = P^ A P . Similar matrices represent the same linear map under two possibly different bases, with being the change-of-basis matrix. A transformation is called a similarity transformation or conjugation of the matrix . In the general linear group, similarity is therefore the same as conjugacy, and similar matrices are also called conjugate; however, in a given subgroup of the general linear group, the notion of conjugacy may be more restrictive than similarity, since it requires that be chosen to lie in . Motivating example When defining a linear transformation, it can be the case that a change of basis can result in a simpler form of the same transformation. For example, the matrix representing a rotation in when the axis of rotation is not aligned with the coordinate axis can be complicated to compute. If the axis of rotation were ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


picture info

Canonical Form
In mathematics and computer science, a canonical, normal, or standard form of a mathematical object is a standard way of presenting that object as a mathematical expression. Often, it is one which provides the simplest representation of an object and allows it to be identified in a unique way. The distinction between "canonical" and "normal" forms varies from subfield to subfield. In most fields, a canonical form specifies a ''unique'' representation for every object, while a normal form simply specifies its form, without the requirement of uniqueness. The canonical form of a positive integer in decimal representation is a finite sequence of digits that does not begin with zero. More generally, for a class of objects on which an equivalence relation is defined, a canonical form consists in the choice of a specific object in each class. For example: *Jordan normal form is a canonical form for matrix similarity. *The row echelon form is a canonical form, when one considers as equ ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


Superdiagonal
In linear algebra, the main diagonal (sometimes principal diagonal, primary diagonal, leading diagonal, major diagonal, or good diagonal) of a matrix A is the list of entries a_ where i = j. All off-diagonal elements are zero in a diagonal matrix. The following four matrices have their main diagonals indicated by red ones: \begin \color & 0 & 0\\ 0 & \color & 0\\ 0 & 0 & \color\end \qquad \begin \color & 0 & 0 & 0 \\ 0 & \color & 0 & 0 \\ 0 & 0 & \color & 0 \end \qquad \begin \color & 0 & 0 \\ 0 & \color & 0 \\ 0 & 0 & \color \\ 0 & 0 & 0 \end \qquad \begin \color & 0 & 0 & 0 \\ 0 & \color & 0 & 0 \\ 0 & 0 & \color & 0 \\ 0 & 0 & 0 & \color \end Square matrices For a square matrix, the ''diagonal'' (or ''main diagonal'' or ''principal diagonal'') is the diagonal line of entries running from the top-left corner to the bottom-right corner. For a matrix A with row index specified by i and column index specified by j, these would be entries A_ with i = j. For example, the i ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


Shift Matrix
In mathematics, a shift matrix is a binary matrix with ones only on the superdiagonal or subdiagonal, and zeroes elsewhere. A shift matrix ''U'' with ones on the superdiagonal is an upper shift matrix. The alternative subdiagonal matrix ''L'' is unsurprisingly known as a lower shift matrix. The (''i'', ''j'')th component of ''U'' and ''L'' are :U_ = \delta_, \quad L_ = \delta_, where \delta_ is the Kronecker delta symbol. For example, the 5 × 5 shift matrices are :U_5 = \begin 0 & 1 & 0 & 0 & 0 \\ 0 & 0 & 1 & 0 & 0 \\ 0 & 0 & 0 & 1 & 0 \\ 0 & 0 & 0 & 0 & 1 \\ 0 & 0 & 0 & 0 & 0 \end \quad L_5 = \begin 0 & 0 & 0 & 0 & 0 \\ 1 & 0 & 0 & 0 & 0 \\ 0 & 1 & 0 & 0 & 0 \\ 0 & 0 & 1 & 0 & 0 \\ 0 & 0 & 0 & 1 & 0 \end. Clearly, the transpose of a lower shift matrix is an upper shift matrix and vice versa. As a linear transformation, a lower shift matrix shifts the components of a column vector one position down, with a zero appearing in the first position. An upper shift matr ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  




Jordan–Chevalley Decomposition
In mathematics, specifically linear algebra, the Jordan–Chevalley decomposition, named after Camille Jordan and Claude Chevalley, expresses a linear operator in a unique way as the sum of two other linear operators which are simpler to understand. Specifically, one part is potentially diagonalisable and the other is nilpotent. The two parts are polynomials in the operator, which makes them behave nicely in algebraic manipulations. The decomposition has a short description when the Jordan normal form of the operator is given, but it exists under weaker hypotheses than are needed for the existence of a Jordan normal form. Hence the Jordan–Chevalley decomposition can be seen as a generalisation of the Jordan normal form, which is also reflected in several proofs of it. It is closely related to the Wedderburn principal theorem about associative algebras, which also leads to several analogues in Lie algebras. Analogues of the Jordan–Chevalley decomposition also exist for elements ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]  


picture info

Diagonalizable Matrix
In linear algebra, a square matrix A is called diagonalizable or non-defective if it is matrix similarity, similar to a diagonal matrix. That is, if there exists an invertible matrix P and a diagonal matrix D such that . This is equivalent to (Such D are not unique.) This property exists for any linear map: for a dimension (vector space), finite-dimensional vector space a linear map T:V\to V is called diagonalizable if there exists an Basis (linear algebra)#Ordered bases and coordinates, ordered basis of V consisting of eigenvectors of T. These definitions are equivalent: if T has a matrix (mathematics), matrix representation A = PDP^ as above, then the column vectors of P form a basis consisting of eigenvectors of and the diagonal entries of D are the corresponding eigenvalues of with respect to this eigenvector basis, T is represented by Diagonalization is the process of finding the above P and and makes many subsequent computations easier. One can raise a diag ...
[...More Info...]      
[...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]