HOME

TheInfoList



OR:

In mathematics, particularly in linear algebra, a skew-symmetric (or antisymmetric or antimetric) matrix is a square matrix whose transpose equals its negative. That is, it satisfies the condition In terms of the entries of the matrix, if a_ denotes the entry in the i-th row and j-th column, then the skew-symmetric condition is equivalent to


Example

The matrix :A = \begin 0 & 2 & -45 \\ -2 & 0 & -4 \\ 45 & 4 & 0 \end is skew-symmetric because : -A = \begin 0 & -2 & 45 \\ 2 & 0 & 4 \\ -45 & -4 & 0 \end = A^\textsf .


Properties

Throughout, we assume that all matrix entries belong to a
field Field may refer to: Expanses of open ground * Field (agriculture), an area of land used for agricultural purposes * Airfield, an aerodrome that lacks the infrastructure of an airport * Battlefield * Lawn, an area of mowed grass * Meadow, a grass ...
\mathbb whose characteristic is not equal to 2. That is, we assume that , where 1 denotes the multiplicative identity and 0 the additive identity of the given field. If the characteristic of the field is 2, then a skew-symmetric matrix is the same thing as a symmetric matrix. * The sum of two skew-symmetric matrices is skew-symmetric. * A scalar multiple of a skew-symmetric matrix is skew-symmetric. * The elements on the diagonal of a skew-symmetric matrix are zero, and therefore its trace equals zero. * If A is a real skew-symmetric matrix and \lambda is a real
eigenvalue In linear algebra, an eigenvector () or characteristic vector of a linear transformation is a nonzero vector that changes at most by a scalar factor when that linear transformation is applied to it. The corresponding eigenvalue, often denoted ...
, then \lambda = 0, i.e. the nonzero eigenvalues of a skew-symmetric matrix are non-real. * If A is a real skew-symmetric matrix, then I + A is invertible, where I is the identity matrix. * If A is a skew-symmetric matrix then A^2 is a symmetric negative semi-definite matrix.


Vector space structure

As a result of the first two properties above, the set of all skew-symmetric matrices of a fixed size forms a vector space. The space of n \times n skew-symmetric matrices has dimension \fracn(n - 1). Let \mbox_n denote the space of n \times n matrices. A skew-symmetric matrix is determined by \fracn(n - 1) scalars (the number of entries above the main diagonal); a symmetric matrix is determined by \fracn(n + 1) scalars (the number of entries on or above the main diagonal). Let \mbox_n denote the space of n \times n skew-symmetric matrices and \mbox_n denote the space of n \times n symmetric matrices. If A \in \mbox_n then A = \frac\left(A - A^\mathsf\right) + \frac\left(A + A^\mathsf\right). Notice that \frac\left(A - A^\textsf\right) \in \mbox_n and \frac\left(A + A^\textsf\right) \in \mbox_n. This is true for every square matrix A with entries from any
field Field may refer to: Expanses of open ground * Field (agriculture), an area of land used for agricultural purposes * Airfield, an aerodrome that lacks the infrastructure of an airport * Battlefield * Lawn, an area of mowed grass * Meadow, a grass ...
whose characteristic is different from 2. Then, since \mbox_n = \mbox_n + \mbox_n and \mbox_n \cap \mbox_n = \, \mbox_n = \mbox_n \oplus \mbox_n, where \oplus denotes the
direct sum The direct sum is an operation between structures in abstract algebra, a branch of mathematics. It is defined differently, but analogously, for different kinds of structures. To see how the direct sum is used in abstract algebra, consider a more ...
. Denote by \langle \cdot, \cdot \rangle the standard inner product on \R^n. The real n \times n matrix A is skew-symmetric if and only if \langle Ax,y \rangle = - \langle x, Ay\rangle \quad \text x, y \in \R^n. This is also equivalent to \langle x, Ax \rangle = 0 for all x \in \R^n (one implication being obvious, the other a plain consequence of \langle x + y, A(x + y)\rangle = 0 for all x and y). Since this definition is independent of the choice of
basis Basis may refer to: Finance and accounting *Adjusted basis, the net cost of an asset after adjusting for various tax-related items *Basis point, 0.01%, often used in the context of interest rates *Basis trading, a trading strategy consisting of ...
, skew-symmetry is a property that depends only on the
linear operator In mathematics, and more specifically in linear algebra, a linear map (also called a linear mapping, linear transformation, vector space homomorphism, or in some contexts linear function) is a mapping V \to W between two vector spaces that pr ...
A and a choice of inner product. 3 \times 3 skew symmetric matrices can be used to represent
cross product In mathematics, the cross product or vector product (occasionally directed area product, to emphasize its geometric significance) is a binary operation on two vectors in a three-dimensional oriented Euclidean vector space (named here E), and is d ...
s as matrix multiplications.


Determinant

Let A be a n \times n skew-symmetric matrix. The determinant of A satisfies :\det(A) = \det\left(A^\textsf\right) = \det(-A) = (-1)^n \det(A). In particular, if n is odd, and since the underlying field is not of characteristic 2, the determinant vanishes. Hence, all odd dimension skew symmetric matrices are singular as their determinants are always zero. This result is called Jacobi’s theorem, after
Carl Gustav Jacobi Carl Gustav Jacob Jacobi (; ; 10 December 1804 – 18 February 1851) was a German mathematician who made fundamental contributions to elliptic functions, Dynamics (mechanics), dynamics, differential equations, determinants, and number theory. H ...
(Eves, 1980). The even-dimensional case is more interesting. It turns out that the determinant of A for n even can be written as the square of a polynomial in the entries of A, which was first proved by Cayley: :\det(A) = \operatorname(A)^2. This polynomial is called the ''
Pfaffian In mathematics, the determinant of a skew-symmetric matrix can always be written as the square of a polynomial in the matrix entries, a polynomial with integer coefficients that only depend on the size of the matrix. The value of this polynomial, ...
'' of A and is denoted \operatorname(A). Thus the determinant of a real skew-symmetric matrix is always non-negative. However this last fact can be proved in an elementary way as follows: the eigenvalues of a real skew-symmetric matrix are purely imaginary (see below) and to every eigenvalue there corresponds the conjugate eigenvalue with the same multiplicity; therefore, as the determinant is the product of the eigenvalues, each one repeated according to its multiplicity, it follows at once that the determinant, if it is not 0, is a positive real number. The number of distinct terms s(n) in the expansion of the determinant of a skew-symmetric matrix of order n has been considered already by Cayley, Sylvester, and Pfaff. Due to cancellations, this number is quite small as compared the number of terms of a generic matrix of order n, which is n!. The sequence s(n) is :1, 0, 1, 0, 6, 0, 120, 0, 5250, 0, 395010, 0, … and it is encoded in the exponential generating function :\sum_^\infty \fracx^n = \left(1 - x^2\right)^\exp\left(\frac\right). The latter yields to the asymptotics (for n even) :s(n) = \pi^ 2^\frac \Gamma\left(\frac\right)\left(\frac\right)^ \left(1 + O\left(\frac\right)\right). The number of positive and negative terms are approximatively a half of the total, although their difference takes larger and larger positive and negative values as n increases .


Cross product

Three-by-three skew-symmetric matrices can be used to represent cross products as matrix multiplications. Consider vectors \mathbf = \left(a_1\ a_2\ a_3\right)^\textsf and \mathbf = \left(b_1\ b_2\ b_3\right)^\textsf. Then, defining the matrix : mathbf = \begin \,\,0 & \!-a_3 & \,\,\,a_2 \\ \,\,\,a_3 & 0 & \!-a_1 \\ \!-a_2 & \,\,a_1 & \,\,0 \end, the cross product can be written as : \mathbf\times\mathbf = mathbf\mathbf. This can be immediately verified by computing both sides of the previous equation and comparing each corresponding element of the results. One actually has : mathbf = mathbf mathbf - mathbf mathbf; i.e., the commutator of skew-symmetric three-by-three matrices can be identified with the cross-product of three-vectors. Since the skew-symmetric three-by-three matrices are the Lie algebra of the rotation group SO(3) this elucidates the relation between three-space \mathbb^3, the cross product and three-dimensional rotations. More on infinitesimal rotations can be found below.


Spectral theory

Since a matrix is similar to its own transpose, they must have the same eigenvalues. It follows that the
eigenvalue In linear algebra, an eigenvector () or characteristic vector of a linear transformation is a nonzero vector that changes at most by a scalar factor when that linear transformation is applied to it. The corresponding eigenvalue, often denoted ...
s of a skew-symmetric matrix always come in pairs ±λ (except in the odd-dimensional case where there is an additional unpaired 0 eigenvalue). From the
spectral theorem In mathematics, particularly linear algebra and functional analysis, a spectral theorem is a result about when a linear operator or matrix can be diagonalized (that is, represented as a diagonal matrix in some basis). This is extremely useful be ...
, for a real skew-symmetric matrix the nonzero eigenvalues are all pure imaginary and thus are of the form \lambda_1 i, -\lambda_1 i, \lambda_2 i, -\lambda_2 i, \ldots where each of the \lambda_k are real. Real skew-symmetric matrices are normal matrices (they commute with their adjoints) and are thus subject to the
spectral theorem In mathematics, particularly linear algebra and functional analysis, a spectral theorem is a result about when a linear operator or matrix can be diagonalized (that is, represented as a diagonal matrix in some basis). This is extremely useful be ...
, which states that any real skew-symmetric matrix can be diagonalized by a
unitary matrix In linear algebra, a complex square matrix is unitary if its conjugate transpose is also its inverse, that is, if U^* U = UU^* = UU^ = I, where is the identity matrix. In physics, especially in quantum mechanics, the conjugate transpose ...
. Since the eigenvalues of a real skew-symmetric matrix are imaginary, it is not possible to diagonalize one by a real matrix. However, it is possible to bring every skew-symmetric matrix to a
block diagonal In mathematics, a block matrix or a partitioned matrix is a matrix that is '' interpreted'' as having been broken into sections called blocks or submatrices. Intuitively, a matrix interpreted as a block matrix can be visualized as the original ma ...
form by a special orthogonal transformation. Specifically, every 2n \times 2n real skew-symmetric matrix can be written in the form A = Q\Sigma Q^\textsf where Q is orthogonal and :\Sigma = \begin \begin0 & \lambda_1 \\ -\lambda_1 & 0\end & 0 & \cdots & 0 \\ 0 & \begin0 & \lambda_2 \\ -\lambda_2 & 0\end & & 0 \\ \vdots & & \ddots & \vdots \\ 0 & 0 & \cdots & \begin0 & \lambda_r\\ -\lambda_r & 0\end \\ & & & & \begin0 \\ & \ddots \\ & & 0 \end \end for real positive-definite \lambda_k. The nonzero eigenvalues of this matrix are ±λ''k'' ''i''. In the odd-dimensional case Σ always has at least one row and column of zeros. More generally, every complex skew-symmetric matrix can be written in the form A = U \Sigma U^ where U is unitary and \Sigma has the block-diagonal form given above with \lambda_k still real positive-definite. This is an example of the Youla decomposition of a complex square matrix.


Skew-symmetric and alternating forms

A skew-symmetric form \varphi on a vector space V over a
field Field may refer to: Expanses of open ground * Field (agriculture), an area of land used for agricultural purposes * Airfield, an aerodrome that lacks the infrastructure of an airport * Battlefield * Lawn, an area of mowed grass * Meadow, a grass ...
K of arbitrary characteristic is defined to be a
bilinear form In mathematics, a bilinear form is a bilinear map on a vector space (the elements of which are called '' vectors'') over a field ''K'' (the elements of which are called ''scalars''). In other words, a bilinear form is a function that is linear ...
:\varphi: V \times V \mapsto K such that for all v, w in V, :\varphi(v, w) = -\varphi(w, v). This defines a form with desirable properties for vector spaces over fields of characteristic not equal to 2, but in a vector space over a field of characteristic 2, the definition is equivalent to that of a symmetric form, as every element is its own additive inverse. Where the vector space V is over a field of arbitrary characteristic including characteristic 2, we may define an alternating form as a bilinear form \varphi such that for all vectors v in V :\varphi(v, v) = 0. This is equivalent to a skew-symmetric form when the field is not of characteristic 2, as seen from :0 = \varphi(v + w, v + w) = \varphi(v, v) + \varphi(v, w) + \varphi(w, v) + \varphi(w, w) = \varphi(v, w) + \varphi(w, v), whence :\varphi(v, w) = -\varphi(w, v). A bilinear form \varphi will be represented by a matrix A such that \varphi(v,w) = v^\textsfAw, once a
basis Basis may refer to: Finance and accounting *Adjusted basis, the net cost of an asset after adjusting for various tax-related items *Basis point, 0.01%, often used in the context of interest rates *Basis trading, a trading strategy consisting of ...
of V is chosen, and conversely an n \times n matrix A on K^n gives rise to a form sending (v, w) to v^\textsfAw. For each of symmetric, skew-symmetric and alternating forms, the representing matrices are symmetric, skew-symmetric and alternating respectively.


Infinitesimal rotations

Skew-symmetric matrices over the field of real numbers form the tangent space to the real orthogonal group O(n) at the identity matrix; formally, the
special orthogonal Lie algebra In mathematics, the orthogonal group in dimension , denoted , is the group of distance-preserving transformations of a Euclidean space of dimension that preserve a fixed point, where the group operation is given by composing transformations. ...
. In this sense, then, skew-symmetric matrices can be thought of as ''infinitesimal rotations''. Another way of saying this is that the space of skew-symmetric matrices forms the Lie algebra o(n) of the
Lie group In mathematics, a Lie group (pronounced ) is a group that is also a differentiable manifold. A manifold is a space that locally resembles Euclidean space, whereas groups define the abstract concept of a binary operation along with the ad ...
O(n). The Lie bracket on this space is given by the
commutator In mathematics, the commutator gives an indication of the extent to which a certain binary operation fails to be commutative. There are different definitions used in group theory and ring theory. Group theory The commutator of two elements, ...
: :
, B The comma is a punctuation mark that appears in several variants in different languages. It has the same shape as an apostrophe or single closing quotation mark () in many typefaces, but it differs from them in being placed on the baseline o ...
= AB - BA.\, It is easy to check that the commutator of two skew-symmetric matrices is again skew-symmetric: : \begin A, B^\textsf &= B^\textsf A^\textsf - A^\textsf B^\textsf \\ &= (-B)(-A) - (-A)(-B) = BA - AB = -
, B The comma is a punctuation mark that appears in several variants in different languages. It has the same shape as an apostrophe or single closing quotation mark () in many typefaces, but it differs from them in being placed on the baseline o ...
\, . \end The
matrix exponential In mathematics, the matrix exponential is a matrix function on square matrices analogous to the ordinary exponential function. It is used to solve systems of linear differential equations. In the theory of Lie groups, the matrix exponential gives ...
of a skew-symmetric matrix A is then an
orthogonal matrix In linear algebra, an orthogonal matrix, or orthonormal matrix, is a real square matrix whose columns and rows are orthonormal vectors. One way to express this is Q^\mathrm Q = Q Q^\mathrm = I, where is the transpose of and is the identity ma ...
R: :R = \exp(A) = \sum_^\infty \frac. The image of the exponential map of a Lie algebra always lies in the connected component of the Lie group that contains the identity element. In the case of the Lie group O(n), this connected component is the special orthogonal group SO(n), consisting of all orthogonal matrices with determinant 1. So R = \exp(A) will have determinant +1. Moreover, since the exponential map of a connected compact Lie group is always surjective, it turns out that ''every'' orthogonal matrix with unit determinant can be written as the exponential of some skew-symmetric matrix. In the particular important case of dimension n=2, the exponential representation for an orthogonal matrix reduces to the well-known polar form of a complex number of unit modulus. Indeed, if n=2, a special orthogonal matrix has the form :\begin a & -b \\ b & \,a \end, with a^2 + b^2 = 1. Therefore, putting a = \cos\theta and b = \sin\theta, it can be written :\begin \cos\,\theta & -\sin\,\theta \\ \sin\,\theta & \,\cos\,\theta \end = \exp\left(\theta\begin 0 & -1 \\ 1 & \,0 \end\right), which corresponds exactly to the polar form \cos \theta + i \sin \theta = e^ of a complex number of unit modulus. The exponential representation of an orthogonal matrix of order n can also be obtained starting from the fact that in dimension n any special orthogonal matrix R can be written as R = QSQ^\textsf, where Q is orthogonal and S is a
block diagonal matrix In mathematics, a block matrix or a partitioned matrix is a matrix that is '' interpreted'' as having been broken into sections called blocks or submatrices. Intuitively, a matrix interpreted as a block matrix can be visualized as the original ma ...
with \lfloor n/2\rfloor blocks of order 2, plus one of order 1 if n is odd; since each single block of order 2 is also an orthogonal matrix, it admits an exponential form. Correspondingly, the matrix ''S'' writes as exponential of a skew-symmetric block matrix \Sigma of the form above, S = \exp(\Sigma), so that R = Q\exp(\Sigma)Q^\textsf = \exp(Q\Sigma Q^\textsf), exponential of the skew-symmetric matrix Q\Sigma Q^\textsf. Conversely, the surjectivity of the exponential map, together with the above-mentioned block-diagonalization for skew-symmetric matrices, implies the block-diagonalization for orthogonal matrices.


Coordinate-free

More intrinsically (i.e., without using coordinates), skew-symmetric linear transformations on a vector space V with an inner product may be defined as the
bivector In mathematics, a bivector or 2-vector is a quantity in exterior algebra or geometric algebra that extends the idea of scalars and vectors. If a scalar is considered a degree-zero quantity, and a vector is a degree-one quantity, then a bivector ca ...
s on the space, which are sums of simple bivectors ( 2-blades) v \wedge w. The correspondence is given by the map v \wedge w \mapsto v^* \otimes w - w^* \otimes v, where v^* is the covector dual to the vector v; in orthonormal coordinates these are exactly the elementary skew-symmetric matrices. This characterization is used in interpreting the
curl cURL (pronounced like "curl", UK: , US: ) is a computer software project providing a library (libcurl) and command-line tool (curl) for transferring data using various network protocols. The name stands for "Client URL". History cURL was fir ...
of a vector field (naturally a 2-vector) as an infinitesimal rotation or "curl", hence the name.


Skew-symmetrizable matrix

An n \times n matrix A is said to be skew-symmetrizable if there exists an invertible diagonal matrix D such that DA is skew-symmetric. For real n \times n matrices, sometimes the condition for D to have positive entries is added.


See also

*
Cayley transform In mathematics, the Cayley transform, named after Arthur Cayley, is any of a cluster of related things. As originally described by , the Cayley transform is a mapping between skew-symmetric matrices and special orthogonal matrices. The transform i ...
* Symmetric matrix *
Skew-Hermitian matrix __NOTOC__ In linear algebra, a square matrix with complex entries is said to be skew-Hermitian or anti-Hermitian if its conjugate transpose is the negative of the original matrix. That is, the matrix A is skew-Hermitian if it satisfies the relatio ...
* Symplectic matrix * Symmetry in mathematics


References


Further reading

* * *


External links

* * *
FortranFortran90
{{Authority control Matrices