HOME

TheInfoList



OR:

In
linear algebra Linear algebra is the branch of mathematics concerning linear equations such as :a_1x_1+\cdots +a_nx_n=b, linear maps such as :(x_1, \ldots, x_n) \mapsto a_1x_1+\cdots +a_nx_n, and their representations in vector spaces and through matrix (mathemat ...
, an orthogonal matrix, or orthonormal matrix, is a real
square matrix In mathematics, a square matrix is a Matrix (mathematics), matrix with the same number of rows and columns. An ''n''-by-''n'' matrix is known as a square matrix of order Any two square matrices of the same order can be added and multiplied. Squ ...
whose columns and rows are
orthonormal In linear algebra, two vectors in an inner product space are orthonormal if they are orthogonal unit vectors. A unit vector means that the vector has a length of 1, which is also known as normalized. Orthogonal means that the vectors are all perpe ...
vectors. One way to express this is Q^\mathrm Q = Q Q^\mathrm = I, where is the
transpose In linear algebra, the transpose of a Matrix (mathematics), matrix is an operator which flips a matrix over its diagonal; that is, it switches the row and column indices of the matrix by producing another matrix, often denoted by (among other ...
of and is the
identity matrix In linear algebra, the identity matrix of size n is the n\times n square matrix with ones on the main diagonal and zeros elsewhere. It has unique properties, for example when the identity matrix represents a geometric transformation, the obje ...
. This leads to the equivalent characterization: a matrix is orthogonal if its transpose is equal to its inverse: Q^\mathrm=Q^, where is the inverse of . An orthogonal matrix is necessarily invertible (with inverse ),
unitary Unitary may refer to: Mathematics * Unitary divisor * Unitary element * Unitary group * Unitary matrix * Unitary morphism * Unitary operator * Unitary transformation * Unitary representation * Unitarity (physics) * ''E''-unitary inverse semigr ...
(), where is the
Hermitian adjoint In mathematics, specifically in operator theory, each linear operator A on an inner product space defines a Hermitian adjoint (or adjoint) operator A^* on that space according to the rule :\langle Ax,y \rangle = \langle x,A^*y \rangle, where \l ...
(
conjugate transpose In mathematics, the conjugate transpose, also known as the Hermitian transpose, of an m \times n complex matrix \mathbf is an n \times m matrix obtained by transposing \mathbf and applying complex conjugation to each entry (the complex conjugate ...
) of , and therefore normal () over the
real number In mathematics, a real number is a number that can be used to measure a continuous one- dimensional quantity such as a duration or temperature. Here, ''continuous'' means that pairs of values can have arbitrarily small differences. Every re ...
s. The
determinant In mathematics, the determinant is a Scalar (mathematics), scalar-valued function (mathematics), function of the entries of a square matrix. The determinant of a matrix is commonly denoted , , or . Its value characterizes some properties of the ...
of any orthogonal matrix is either +1 or −1. As a
linear transformation In mathematics, and more specifically in linear algebra, a linear map (also called a linear mapping, linear transformation, vector space homomorphism, or in some contexts linear function) is a mapping V \to W between two vector spaces that pr ...
, an orthogonal matrix preserves the
inner product In mathematics, an inner product space (or, rarely, a Hausdorff pre-Hilbert space) is a real vector space or a complex vector space with an operation called an inner product. The inner product of two vectors in the space is a scalar, ofte ...
of vectors, and therefore acts as an
isometry In mathematics, an isometry (or congruence, or congruent transformation) is a distance-preserving transformation between metric spaces, usually assumed to be bijective. The word isometry is derived from the Ancient Greek: ἴσος ''isos'' me ...
of
Euclidean space Euclidean space is the fundamental space of geometry, intended to represent physical space. Originally, in Euclid's ''Elements'', it was the three-dimensional space of Euclidean geometry, but in modern mathematics there are ''Euclidean spaces ...
, such as a
rotation Rotation or rotational/rotary motion is the circular movement of an object around a central line, known as an ''axis of rotation''. A plane figure can rotate in either a clockwise or counterclockwise sense around a perpendicular axis intersect ...
, reflection or
rotoreflection In geometry, an improper rotation. (also called rotation-reflection, rotoreflection, rotary reflection,. or rotoinversion) is an isometry in Euclidean space that is a combination of a Rotation (geometry), rotation about an axis and a reflection ( ...
. In other words, it is a
unitary transformation In mathematics, a unitary transformation is a linear isomorphism that preserves the inner product: the inner product of two vectors before the transformation is equal to their inner product after the transformation. Formal definition More precise ...
. The set of orthogonal matrices, under multiplication, forms the
group A group is a number of persons or things that are located, gathered, or classed together. Groups of people * Cultural group, a group whose members share the same cultural identity * Ethnic group, a group whose members share the same ethnic iden ...
, known as the
orthogonal group In mathematics, the orthogonal group in dimension , denoted , is the Group (mathematics), group of isometry, distance-preserving transformations of a Euclidean space of dimension that preserve a fixed point, where the group operation is given by ...
. The
subgroup In group theory, a branch of mathematics, a subset of a group G is a subgroup of G if the members of that subset form a group with respect to the group operation in G. Formally, given a group (mathematics), group under a binary operation  ...
consisting of orthogonal matrices with determinant +1 is called the
special orthogonal group In mathematics, the orthogonal group in dimension , denoted , is the group of distance-preserving transformations of a Euclidean space of dimension that preserve a fixed point, where the group operation is given by composing transformations. ...
, and each of its elements is a special orthogonal matrix. As a linear transformation, every special orthogonal matrix acts as a rotation.


Overview

An orthogonal matrix is the real specialization of a unitary matrix, and thus always a
normal matrix In mathematics, a complex square matrix is normal if it commutes with its conjugate transpose : :A \text \iff A^*A = AA^* . The concept of normal matrices can be extended to normal operators on infinite-dimensional normed spaces and to nor ...
. Although we consider only real matrices here, the definition can be used for matrices with entries from any field. However, orthogonal matrices arise naturally from
dot product In mathematics, the dot product or scalar productThe term ''scalar product'' means literally "product with a Scalar (mathematics), scalar as a result". It is also used for other symmetric bilinear forms, for example in a pseudo-Euclidean space. N ...
s, and for matrices of complex numbers that leads instead to the unitary requirement. Orthogonal matrices preserve the dot product, so, for vectors and in an -dimensional real
Euclidean space Euclidean space is the fundamental space of geometry, intended to represent physical space. Originally, in Euclid's ''Elements'', it was the three-dimensional space of Euclidean geometry, but in modern mathematics there are ''Euclidean spaces ...
\cdot = \left(Q \right) \cdot \left(Q \right) where is an orthogonal matrix. To see the inner product connection, consider a vector in an -dimensional real
Euclidean space Euclidean space is the fundamental space of geometry, intended to represent physical space. Originally, in Euclid's ''Elements'', it was the three-dimensional space of Euclidean geometry, but in modern mathematics there are ''Euclidean spaces ...
. Written with respect to an orthonormal basis, the squared length of is . If a linear transformation, in matrix form , preserves vector lengths, then ^\mathrm = (Q)^\mathrm(Q) = ^\mathrm Q^\mathrm Q . Thus
finite-dimensional In mathematics, the dimension of a vector space ''V'' is the cardinality (i.e., the number of vectors) of a basis of ''V'' over its base field. p. 44, §2.36 It is sometimes called Hamel dimension (after Georg Hamel) or algebraic dimension to d ...
linear isometries—rotations, reflections, and their combinations—produce orthogonal matrices. The converse is also true: orthogonal matrices imply orthogonal transformations. However, linear algebra includes orthogonal transformations between spaces which may be neither finite-dimensional nor of the same dimension, and these have no orthogonal matrix equivalent. Orthogonal matrices are important for a number of reasons, both theoretical and practical. The orthogonal matrices form a
group A group is a number of persons or things that are located, gathered, or classed together. Groups of people * Cultural group, a group whose members share the same cultural identity * Ethnic group, a group whose members share the same ethnic iden ...
under matrix multiplication, the
orthogonal group In mathematics, the orthogonal group in dimension , denoted , is the Group (mathematics), group of isometry, distance-preserving transformations of a Euclidean space of dimension that preserve a fixed point, where the group operation is given by ...
denoted by , which—with its subgroups—is widely used in mathematics and the physical sciences. For example, the
point group In geometry, a point group is a group (mathematics), mathematical group of symmetry operations (isometry, isometries in a Euclidean space) that have a Fixed point (mathematics), fixed point in common. The Origin (mathematics), coordinate origin o ...
of a molecule is a subgroup of O(3). Because floating point versions of orthogonal matrices have advantageous properties, they are key to many algorithms in numerical linear algebra, such as decomposition. As another example, with appropriate normalization the
discrete cosine transform A discrete cosine transform (DCT) expresses a finite sequence of data points in terms of a sum of cosine functions oscillating at different frequency, frequencies. The DCT, first proposed by Nasir Ahmed (engineer), Nasir Ahmed in 1972, is a widely ...
(used in MP3 compression) is represented by an orthogonal matrix.


Examples

Below are a few examples of small orthogonal matrices and possible interpretations. * \begin 1 & 0 \\ 0 & 1 \\ \end    (identity transformation) * \begin \cos \theta & -\sin \theta \\ \sin \theta & \cos \theta \\ \end    (rotation about the origin) * \begin 1 & 0 \\ 0 & -1 \\ \end    (reflection across ''x''-axis) * \begin 0 & 0 & 0 & 1 \\ 0 & 0 & 1 & 0 \\ 1 & 0 & 0 & 0 \\ 0 & 1 & 0 & 0 \end    (permutation of coordinate axes)


Elementary constructions


Lower dimensions

The simplest orthogonal matrices are the matrices and ��1 which we can interpret as the identity and a reflection of the real line across the origin. The matrices have the form \begin p & t\\ q & u \end, which orthogonality demands satisfy the three equations \begin 1 & = p^2+t^2, \\ 1 & = q^2+u^2, \\ 0 & = pq+tu. \end In consideration of the first equation, without loss of generality let , ; then either , or , . We can interpret the first case as a rotation by (where is the identity), and the second as a reflection across a line at an angle of . \begin \cos \theta & -\sin \theta \\ \sin \theta & \cos \theta \\ \end\text\qquad \begin \cos \theta & \sin \theta \\ \sin \theta & -\cos \theta \\ \end\text The special case of the reflection matrix with generates a reflection about the line at 45° given by and therefore exchanges and ; it is a
permutation matrix In mathematics, particularly in matrix theory, a permutation matrix is a square binary matrix that has exactly one entry of 1 in each row and each column with all other entries 0. An permutation matrix can represent a permutation of elements. ...
, with a single 1 in each column and row (and otherwise 0): \begin 0 & 1\\ 1 & 0 \end. The identity is also a permutation matrix. A reflection is its own inverse, which implies that a reflection matrix is
symmetric Symmetry () in everyday life refers to a sense of harmonious and beautiful proportion and balance. In mathematics, the term has a more precise definition and is usually used to refer to an object that is invariant under some transformations ...
(equal to its transpose) as well as orthogonal. The product of two rotation matrices is a
rotation matrix In linear algebra, a rotation matrix is a transformation matrix that is used to perform a rotation (mathematics), rotation in Euclidean space. For example, using the convention below, the matrix :R = \begin \cos \theta & -\sin \theta \\ \sin \t ...
, and the product of two reflection matrices is also a rotation matrix.


Higher dimensions

Regardless of the dimension, it is always possible to classify orthogonal matrices as purely rotational or not, but for matrices and larger the non-rotational matrices can be more complicated than reflections. For example, \begin -1 & 0 & 0\\ 0 & -1 & 0\\ 0 & 0 & -1 \end\text \begin 0 & -1 & 0\\ 1 & 0 & 0\\ 0 & 0 & -1 \end represent an '' inversion'' through the origin and a '' rotoinversion'', respectively, about the -axis. Rotations become more complicated in higher dimensions; they can no longer be completely characterized by one angle, and may affect more than one planar subspace. It is common to describe a rotation matrix in terms of an axis and angle, but this only works in three dimensions. Above three dimensions two or more angles are needed, each associated with a plane of rotation. However, we have elementary building blocks for permutations, reflections, and rotations that apply in general.


Primitives

The most elementary permutation is a transposition, obtained from the identity matrix by exchanging two rows. Any permutation matrix can be constructed as a product of no more than transpositions. A Householder reflection is constructed from a non-null vector as Q = I - 2 \frac . Here the numerator is a symmetric matrix while the denominator is a number, the squared magnitude of . This is a reflection in the hyperplane perpendicular to (negating any vector component parallel to ). If is a unit vector, then suffices. A Householder reflection is typically used to simultaneously zero the lower part of a column. Any orthogonal matrix of size can be constructed as a product of at most such reflections. A Givens rotation acts on a two-dimensional (planar) subspace spanned by two coordinate axes, rotating by a chosen angle. It is typically used to zero a single subdiagonal entry. Any rotation matrix of size can be constructed as a product of at most such rotations. In the case of matrices, three such rotations suffice; and by fixing the sequence we can thus describe all rotation matrices (though not uniquely) in terms of the three angles used, often called
Euler angles The Euler angles are three angles introduced by Leonhard Euler to describe the Orientation (geometry), orientation of a rigid body with respect to a fixed coordinate system.Novi Commentarii academiae scientiarum Petropolitanae 20, 1776, pp. 189� ...
. A Jacobi rotation has the same form as a Givens rotation, but is used to zero both off-diagonal entries of a symmetric submatrix.


Properties


Matrix properties

A real square matrix is orthogonal
if and only if In logic and related fields such as mathematics and philosophy, "if and only if" (often shortened as "iff") is paraphrased by the biconditional, a logical connective between statements. The biconditional is true in two cases, where either bo ...
its columns form an
orthonormal basis In mathematics, particularly linear algebra, an orthonormal basis for an inner product space V with finite Dimension (linear algebra), dimension is a Basis (linear algebra), basis for V whose vectors are orthonormal, that is, they are all unit vec ...
of the
Euclidean space Euclidean space is the fundamental space of geometry, intended to represent physical space. Originally, in Euclid's ''Elements'', it was the three-dimensional space of Euclidean geometry, but in modern mathematics there are ''Euclidean spaces ...
with the ordinary Euclidean
dot product In mathematics, the dot product or scalar productThe term ''scalar product'' means literally "product with a Scalar (mathematics), scalar as a result". It is also used for other symmetric bilinear forms, for example in a pseudo-Euclidean space. N ...
, which is the case if and only if its rows form an orthonormal basis of . It might be tempting to suppose a matrix with orthogonal (not orthonormal) columns would be called an orthogonal matrix, but such matrices have no special interest and no special name; they only satisfy , with a
diagonal matrix In linear algebra, a diagonal matrix is a matrix in which the entries outside the main diagonal are all zero; the term usually refers to square matrices. Elements of the main diagonal can either be zero or nonzero. An example of a 2×2 diagon ...
. The
determinant In mathematics, the determinant is a Scalar (mathematics), scalar-valued function (mathematics), function of the entries of a square matrix. The determinant of a matrix is commonly denoted , , or . Its value characterizes some properties of the ...
of any orthogonal matrix is +1 or −1. This follows from basic facts about determinants, as follows: 1=\det(I)=\det\left(Q^\mathrmQ\right)=\det\left(Q^\mathrm\right)\det(Q)=\bigl(\det(Q)\bigr)^2 . The converse is not true; having a determinant of ±1 is no guarantee of orthogonality, even with orthogonal columns, as shown by the following counterexample. \begin 2 & 0 \\ 0 & \frac \end With permutation matrices the determinant matches the
signature A signature (; from , "to sign") is a depiction of someone's name, nickname, or even a simple "X" or other mark that a person writes on documents as a proof of identity and intent. Signatures are often, but not always, Handwriting, handwritt ...
, being +1 or −1 as the parity of the permutation is even or odd, for the determinant is an alternating function of the rows. Stronger than the determinant restriction is the fact that an orthogonal matrix can always be diagonalized over the
complex number In mathematics, a complex number is an element of a number system that extends the real numbers with a specific element denoted , called the imaginary unit and satisfying the equation i^= -1; every complex number can be expressed in the for ...
s to exhibit a full set of
eigenvalues In linear algebra, an eigenvector ( ) or characteristic vector is a vector that has its direction unchanged (or reversed) by a given linear transformation. More precisely, an eigenvector \mathbf v of a linear transformation T is scaled by a ...
, all of which must have (complex) modulus 1.


Group properties

The inverse of every orthogonal matrix is again orthogonal, as is the matrix product of two orthogonal matrices. In fact, the set of all orthogonal matrices satisfies all the axioms of a
group A group is a number of persons or things that are located, gathered, or classed together. Groups of people * Cultural group, a group whose members share the same cultural identity * Ethnic group, a group whose members share the same ethnic iden ...
. It is a
compact Compact as used in politics may refer broadly to a pact or treaty; in more specific cases it may refer to: * Interstate compact, a type of agreement used by U.S. states * Blood compact, an ancient ritual of the Philippines * Compact government, a t ...
Lie group In mathematics, a Lie group (pronounced ) is a group (mathematics), group that is also a differentiable manifold, such that group multiplication and taking inverses are both differentiable. A manifold is a space that locally resembles Eucli ...
of dimension , called the
orthogonal group In mathematics, the orthogonal group in dimension , denoted , is the Group (mathematics), group of isometry, distance-preserving transformations of a Euclidean space of dimension that preserve a fixed point, where the group operation is given by ...
and denoted by . The orthogonal matrices whose determinant is +1 form a
path-connected In topology and related branches of mathematics, a connected space is a topological space that cannot be represented as the union of two or more disjoint non-empty open subsets. Connectedness is one of the principal topological properties t ...
normal subgroup In abstract algebra, a normal subgroup (also known as an invariant subgroup or self-conjugate subgroup) is a subgroup that is invariant under conjugation by members of the group of which it is a part. In other words, a subgroup N of the group ...
of of
index Index (: indexes or indices) may refer to: Arts, entertainment, and media Fictional entities * Index (''A Certain Magical Index''), a character in the light novel series ''A Certain Magical Index'' * The Index, an item on the Halo Array in the ...
2, the
special orthogonal group In mathematics, the orthogonal group in dimension , denoted , is the group of distance-preserving transformations of a Euclidean space of dimension that preserve a fixed point, where the group operation is given by composing transformations. ...
of rotations. The
quotient group A quotient group or factor group is a mathematical group obtained by aggregating similar elements of a larger group using an equivalence relation that preserves some of the group structure (the rest of the structure is "factored out"). For ex ...
is isomorphic to , with the projection map choosing 1or ��1according to the determinant. Orthogonal matrices with determinant −1 do not include the identity, and so do not form a subgroup but only a
coset In mathematics, specifically group theory, a subgroup of a group may be used to decompose the underlying set of into disjoint, equal-size subsets called cosets. There are ''left cosets'' and ''right cosets''. Cosets (both left and right) ...
; it is also (separately) connected. Thus each orthogonal group falls into two pieces; and because the projection map splits, is a
semidirect product In mathematics, specifically in group theory, the concept of a semidirect product is a generalization of a direct product. It is usually denoted with the symbol . There are two closely related concepts of semidirect product: * an ''inner'' sem ...
of by . In practical terms, a comparable statement is that any orthogonal matrix can be produced by taking a rotation matrix and possibly negating one of its columns, as we saw with matrices. If is odd, then the semidirect product is in fact a
direct product In mathematics, a direct product of objects already known can often be defined by giving a new one. That induces a structure on the Cartesian product of the underlying sets from that of the contributing objects. The categorical product is an abs ...
, and any orthogonal matrix can be produced by taking a rotation matrix and possibly negating all of its columns. This follows from the property of determinants that negating a column negates the determinant, and thus negating an odd (but not even) number of columns negates the determinant. Now consider orthogonal matrices with bottom right entry equal to 1. The remainder of the last column (and last row) must be zeros, and the product of any two such matrices has the same form. The rest of the matrix is an orthogonal matrix; thus is a subgroup of (and of all higher groups). \begin & & & 0\\ & \mathrm(n) & & \vdots\\ & & & 0\\ 0 & \cdots & 0 & 1 \end Since an elementary reflection in the form of a Householder matrix can reduce any orthogonal matrix to this constrained form, a series of such reflections can bring any orthogonal matrix to the identity; thus an orthogonal group is a reflection group. The last column can be fixed to any unit vector, and each choice gives a different copy of in ; in this way is a bundle over the unit sphere with fiber . Similarly, is a subgroup of ; and any special orthogonal matrix can be generated by Givens plane rotations using an analogous procedure. The bundle structure persists: . A single rotation can produce a zero in the first row of the last column, and series of rotations will zero all but the last row of the last column of an rotation matrix. Since the planes are fixed, each rotation has only one degree of freedom, its angle. By induction, therefore has (n-1) + (n-2) + \cdots + 1 = \frac degrees of freedom, and so does . Permutation matrices are simpler still; they form, not a Lie group, but only a finite group, the order
symmetric group In abstract algebra, the symmetric group defined over any set is the group whose elements are all the bijections from the set to itself, and whose group operation is the composition of functions. In particular, the finite symmetric grou ...
. By the same kind of argument, is a subgroup of . The even permutations produce the subgroup of permutation matrices of determinant +1, the order
alternating group In mathematics, an alternating group is the Group (mathematics), group of even permutations of a finite set. The alternating group on a set of elements is called the alternating group of degree , or the alternating group on letters and denoted ...
.


Canonical form

More broadly, the effect of any orthogonal matrix separates into independent actions on orthogonal two-dimensional subspaces. That is, if is special orthogonal then one can always find an orthogonal matrix , a (rotational)
change of basis In mathematics, an ordered basis of a vector space of finite dimension allows representing uniquely any element of the vector space by a coordinate vector, which is a sequence of scalars called coordinates. If two different bases are conside ...
, that brings into block diagonal form: P^\mathrmQP = \begin R_1 & & \\ & \ddots & \\ & & R_k \end\ (n\text), \ P^\mathrmQP = \begin R_1 & & & \\ & \ddots & & \\ & & R_k & \\ & & & 1 \end\ (n\text). where the matrices are rotation matrices, and with the remaining entries zero. Exceptionally, a rotation block may be diagonal, . Thus, negating one column if necessary, and noting that a reflection diagonalizes to a +1 and −1, any orthogonal matrix can be brought to the form P^\mathrmQP = \begin \beginR_1 & & \\ & \ddots & \\ & & R_k\end & 0 \\ 0 & \begin\pm 1 & & \\ & \ddots & \\ & & \pm 1\end \\ \end, The matrices give conjugate pairs of eigenvalues lying on the unit circle in the
complex plane In mathematics, the complex plane is the plane (geometry), plane formed by the complex numbers, with a Cartesian coordinate system such that the horizontal -axis, called the real axis, is formed by the real numbers, and the vertical -axis, call ...
; so this decomposition confirms that all
eigenvalues In linear algebra, an eigenvector ( ) or characteristic vector is a vector that has its direction unchanged (or reversed) by a given linear transformation. More precisely, an eigenvector \mathbf v of a linear transformation T is scaled by a ...
have
absolute value In mathematics, the absolute value or modulus of a real number x, is the non-negative value without regard to its sign. Namely, , x, =x if x is a positive number, and , x, =-x if x is negative (in which case negating x makes -x positive), ...
1. If is odd, there is at least one real eigenvalue, +1 or −1; for a rotation, the eigenvector associated with +1 is the rotation axis.


Lie algebra

Suppose the entries of are differentiable functions of , and that gives . Differentiating the orthogonality condition Q^\mathrm Q = I yields \dot^\mathrm Q + Q^\mathrm \dot = 0 Evaluation at () then implies \dot^\mathrm = -\dot . In Lie group terms, this means that the
Lie algebra In mathematics, a Lie algebra (pronounced ) is a vector space \mathfrak g together with an operation called the Lie bracket, an alternating bilinear map \mathfrak g \times \mathfrak g \rightarrow \mathfrak g, that satisfies the Jacobi ident ...
of an orthogonal matrix group consists of skew-symmetric matrices. Going the other direction, the
matrix exponential In mathematics, the matrix exponential is a matrix function on square matrix, square matrices analogous to the ordinary exponential function. It is used to solve systems of linear differential equations. In the theory of Lie groups, the matrix exp ...
of any skew-symmetric matrix is an orthogonal matrix (in fact, special orthogonal). For example, the three-dimensional object physics calls
angular velocity In physics, angular velocity (symbol or \vec, the lowercase Greek letter omega), also known as the angular frequency vector,(UP1) is a pseudovector representation of how the angular position or orientation of an object changes with time, i ...
is a differential rotation, thus a vector in the Lie algebra \mathfrak(3) tangent to . Given , with being a unit vector, the correct skew-symmetric matrix form of is \Omega = \begin 0 & -z\theta & y\theta \\ z\theta & 0 & -x\theta \\ -y\theta & x\theta & 0 \end . The exponential of this is the orthogonal matrix for rotation around axis by angle ; setting , , \exp(\Omega) = \begin 1 - 2s^2 + 2x^2 s^2 & 2xy s^2 - 2z sc & 2xz s^2 + 2y sc\\ 2xy s^2 + 2z sc & 1 - 2s^2 + 2y^2 s^2 & 2yz s^2 - 2x sc\\ 2xz s^2 - 2y sc & 2yz s^2 + 2x sc & 1 - 2s^2 + 2z^2 s^2 \end.


Numerical linear algebra


Benefits

Numerical analysis Numerical analysis is the study of algorithms that use numerical approximation (as opposed to symbolic computation, symbolic manipulations) for the problems of mathematical analysis (as distinguished from discrete mathematics). It is the study of ...
takes advantage of many of the properties of orthogonal matrices for numerical linear algebra, and they arise naturally. For example, it is often desirable to compute an orthonormal basis for a space, or an orthogonal change of bases; both take the form of orthogonal matrices. Having determinant ±1 and all eigenvalues of magnitude 1 is of great benefit for numeric stability. One implication is that the condition number is 1 (which is the minimum), so errors are not magnified when multiplying with an orthogonal matrix. Many algorithms use orthogonal matrices like Householder reflections and Givens rotations for this reason. It is also helpful that, not only is an orthogonal matrix invertible, but its inverse is available essentially free, by exchanging indices. Permutations are essential to the success of many algorithms, including the workhorse
Gaussian elimination In mathematics, Gaussian elimination, also known as row reduction, is an algorithm for solving systems of linear equations. It consists of a sequence of row-wise operations performed on the corresponding matrix of coefficients. This method can a ...
with partial pivoting (where permutations do the pivoting). However, they rarely appear explicitly as matrices; their special form allows more efficient representation, such as a list of indices. Likewise, algorithms using Householder and Givens matrices typically use specialized methods of multiplication and storage. For example, a Givens rotation affects only two rows of a matrix it multiplies, changing a full
multiplication Multiplication is one of the four elementary mathematical operations of arithmetic, with the other ones being addition, subtraction, and division (mathematics), division. The result of a multiplication operation is called a ''Product (mathem ...
of order to a much more efficient order . When uses of these reflections and rotations introduce zeros in a matrix, the space vacated is enough to store sufficient data to reproduce the transform, and to do so robustly. (Following , we do ''not'' store a rotation angle, which is both expensive and badly behaved.)


Decompositions

A number of important matrix decompositions involve orthogonal matrices, including especially: ; decomposition : , orthogonal, upper triangular ;
Singular value decomposition In linear algebra, the singular value decomposition (SVD) is a Matrix decomposition, factorization of a real number, real or complex number, complex matrix (mathematics), matrix into a rotation, followed by a rescaling followed by another rota ...
: , and orthogonal, diagonal matrix ; Eigendecomposition of a symmetric matrix (decomposition according to the
spectral theorem In linear algebra and functional analysis, a spectral theorem is a result about when a linear operator or matrix can be diagonalized (that is, represented as a diagonal matrix in some basis). This is extremely useful because computations involvin ...
) : , symmetric, orthogonal, diagonal ;
Polar decomposition In mathematics, the polar decomposition of a square real or complex matrix A is a factorization of the form A = U P, where U is a unitary matrix, and P is a positive semi-definite Hermitian matrix (U is an orthogonal matrix, and P is a posit ...
: , orthogonal, symmetric positive-semidefinite


Examples

Consider an overdetermined system of linear equations, as might occur with repeated measurements of a physical phenomenon to compensate for experimental errors. Write , where is , . A decomposition reduces to upper triangular . For example, if is then has the form R = \begin \cdot & \cdot & \cdot \\ 0 & \cdot & \cdot \\ 0 & 0 & \cdot \\ 0 & 0 & 0 \\ 0 & 0 & 0 \end. The
linear least squares Linear least squares (LLS) is the least squares approximation of linear functions to data. It is a set of formulations for solving statistical problems involved in linear regression, including variants for ordinary (unweighted), weighted, and ...
problem is to find the that minimizes , which is equivalent to projecting to the subspace spanned by the columns of . Assuming the columns of (and hence ) are independent, the projection solution is found from . Now is square () and invertible, and also equal to . But the lower rows of zeros in are superfluous in the product, which is thus already in lower-triangular upper-triangular factored form, as in
Gaussian elimination In mathematics, Gaussian elimination, also known as row reduction, is an algorithm for solving systems of linear equations. It consists of a sequence of row-wise operations performed on the corresponding matrix of coefficients. This method can a ...
(
Cholesky decomposition In linear algebra, the Cholesky decomposition or Cholesky factorization (pronounced ) is a decomposition of a Hermitian, positive-definite matrix into the product of a lower triangular matrix and its conjugate transpose, which is useful for eff ...
). Here orthogonality is important not only for reducing to , but also for allowing solution without magnifying numerical problems. In the case of a linear system which is underdetermined, or an otherwise non-
invertible matrix In linear algebra, an invertible matrix (''non-singular'', ''non-degenarate'' or ''regular'') is a square matrix that has an inverse. In other words, if some other matrix is multiplied by the invertible matrix, the result can be multiplied by a ...
, singular value decomposition (SVD) is equally useful. With factored as , a satisfactory solution uses the Moore-Penrose
pseudoinverse In mathematics, and in particular, algebra, a generalized inverse (or, g-inverse) of an element ''x'' is an element ''y'' that has some properties of an inverse element but not necessarily all of them. The purpose of constructing a generalized inv ...
, , where merely replaces each non-zero diagonal entry with its reciprocal. Set to . The case of a square invertible matrix also holds interest. Suppose, for example, that is a rotation matrix which has been computed as the composition of numerous twists and turns. Floating point does not match the mathematical ideal of real numbers, so has gradually lost its true orthogonality. A
Gram–Schmidt process In mathematics, particularly linear algebra and numerical analysis, the Gram–Schmidt process or Gram-Schmidt algorithm is a way of finding a set of two or more vectors that are perpendicular to each other. By technical definition, it is a metho ...
could orthogonalize the columns, but it is not the most reliable, nor the most efficient, nor the most invariant method. The
polar decomposition In mathematics, the polar decomposition of a square real or complex matrix A is a factorization of the form A = U P, where U is a unitary matrix, and P is a positive semi-definite Hermitian matrix (U is an orthogonal matrix, and P is a posit ...
factors a matrix into a pair, one of which is the unique ''closest'' orthogonal matrix to the given matrix, or one of the closest if the given matrix is singular. (Closeness can be measured by any
matrix norm In the field of mathematics, norms are defined for elements within a vector space. Specifically, when the vector space comprises matrices, such norms are referred to as matrix norms. Matrix norms differ from vector norms in that they must also ...
invariant under an orthogonal change of basis, such as the spectral norm or the Frobenius norm.) For a near-orthogonal matrix, rapid convergence to the orthogonal factor can be achieved by a "
Newton's method In numerical analysis, the Newton–Raphson method, also known simply as Newton's method, named after Isaac Newton and Joseph Raphson, is a root-finding algorithm which produces successively better approximations to the roots (or zeroes) of a ...
" approach due to (
1990 Important events of 1990 include the Reunification of Germany and the unification of Yemen, the formal beginning of the Human Genome Project (finished in 2003), the launch of the Hubble Space Telescope, the separation of Namibia from South ...
), repeatedly averaging the matrix with its inverse transpose. has published an accelerated method with a convenient convergence test. For example, consider a non-orthogonal matrix for which the simple averaging algorithm takes seven steps \begin3 & 1\\7 & 5\end \rightarrow \begin1.8125 & 0.0625\\3.4375 & 2.6875\end \rightarrow \cdots \rightarrow \begin0.8 & -0.6\\0.6 & 0.8\end and which acceleration trims to two steps (with = 0.353553, 0.565685). \begin3 & 1\\7 & 5\end \rightarrow \begin1.41421 & -1.06066\\1.06066 & 1.41421\end \rightarrow \begin0.8 & -0.6\\0.6 & 0.8\end Gram-Schmidt yields an inferior solution, shown by a Frobenius distance of 8.28659 instead of the minimum 8.12404. \begin3 & 1\\7 & 5\end \rightarrow \begin0.393919 & -0.919145\\0.919145 & 0.393919\end


Randomization

Some numerical applications, such as
Monte Carlo method Monte Carlo methods, or Monte Carlo experiments, are a broad class of computational algorithms that rely on repeated random sampling to obtain numerical results. The underlying concept is to use randomness to solve problems that might be ...
s and exploration of high-dimensional data spaces, require generation of uniformly distributed random orthogonal matrices. In this context, "uniform" is defined in terms of
Haar measure In mathematical analysis, the Haar measure assigns an "invariant volume" to subsets of locally compact topological groups, consequently defining an integral for functions on those groups. This Measure (mathematics), measure was introduced by Alfr� ...
, which essentially requires that the distribution not change if multiplied by any freely chosen orthogonal matrix. Orthogonalizing matrices with
independent Independent or Independents may refer to: Arts, entertainment, and media Artist groups * Independents (artist group), a group of modernist painters based in Pennsylvania, United States * Independentes (English: Independents), a Portuguese artist ...
uniformly distributed random entries does not result in uniformly distributed orthogonal matrices, but the decomposition of independent
normally distributed In probability theory and statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real number, real-valued random variable. The general form of its probability density function is f(x ...
random entries does, as long as the diagonal of contains only positive entries . replaced this with a more efficient idea that later generalized as the "subgroup algorithm" (in which form it works just as well for permutations and rotations). To generate an orthogonal matrix, take an one and a uniformly distributed unit vector of dimension . Construct a Householder reflection from the vector, then apply it to the smaller matrix (embedded in the larger size with a 1 at the bottom right corner).


Nearest orthogonal matrix

The problem of finding the orthogonal matrix nearest a given matrix is related to the Orthogonal Procrustes problem. There are several different ways to get the unique solution, the simplest of which is taking the
singular value decomposition In linear algebra, the singular value decomposition (SVD) is a Matrix decomposition, factorization of a real number, real or complex number, complex matrix (mathematics), matrix into a rotation, followed by a rescaling followed by another rota ...
of and replacing the singular values with ones. Another method expresses the explicitly but requires the use of a
matrix square root In mathematics, the square root of a matrix extends the notion of square root from numbers to Matrix (mathematics), matrices. A matrix is said to be a square root of if the matrix product is equal to . Some authors use the name ''square root' ...
: Q = M \left(M^\mathrm M\right)^ This may be combined with the Babylonian method for extracting the square root of a matrix to give a recurrence which converges to an orthogonal matrix quadratically: Q_ = 2 M \left(Q_n^ M + M^\mathrm Q_n\right)^ where . These iterations are stable provided the condition number of is less than three."Newton's Method for the Matrix Square Root"
, Nicholas J. Higham, Mathematics of Computation, Volume 46, Number 174, 1986. Using a first-order approximation of the inverse and the same initialization results in the modified iteration: N_ = Q_n^\mathrm Q_n P_ = \frac 1 2 Q_n N_ Q_ = 2 Q_n + P_n N_n - 3 P_n


Spin and pin

A subtle technical problem afflicts some uses of orthogonal matrices. Not only are the group components with determinant +1 and −1 not connected to each other, even the +1 component, , is not
simply connected In topology, a topological space is called simply connected (or 1-connected, or 1-simply connected) if it is path-connected and every Path (topology), path between two points can be continuously transformed into any other such path while preserving ...
(except for SO(1), which is trivial). Thus it is sometimes advantageous, or even necessary, to work with a
covering group In mathematics, a covering group of a topological group ''H'' is a covering space ''G'' of ''H'' such that ''G'' is a topological group and the covering map is a continuous (topology), continuous group homomorphism. The map ''p'' is called the c ...
of SO(''n''), the
spin group In mathematics the spin group, denoted Spin(''n''), page 15 is a Lie group whose underlying manifold is the double cover of the special orthogonal group , such that there exists a short exact sequence of Lie groups (when ) :1 \to \mathbb_2 \to \o ...
, . Likewise, has covering groups, the
pin group In mathematics, the pin group is a certain subgroup of the Clifford algebra associated to a quadratic space. It maps 2-to-1 to the orthogonal group, just as the spin group maps 2-to-1 to the special orthogonal group. In general the map from th ...
s, Pin(''n''). For , is simply connected and thus the universal covering group for . By far the most famous example of a spin group is , which is nothing but , or the group of unit
quaternion In mathematics, the quaternion number system extends the complex numbers. Quaternions were first described by the Irish mathematician William Rowan Hamilton in 1843 and applied to mechanics in three-dimensional space. The algebra of quater ...
s. The Pin and Spin groups are found within
Clifford algebra In mathematics, a Clifford algebra is an algebra generated by a vector space with a quadratic form, and is a unital associative algebra with the additional structure of a distinguished subspace. As -algebras, they generalize the real number ...
s, which themselves can be built from orthogonal matrices.


Rectangular matrices

If is not a square matrix, then the conditions and are not equivalent. The condition says that the columns of ''Q'' are orthonormal. This can only happen if is an matrix with (due to linear dependence). Similarly, says that the rows of are orthonormal, which requires . There is no standard terminology for these matrices. They are variously called "semi-orthogonal matrices", "orthonormal matrices", "orthogonal matrices", and sometimes simply "matrices with orthonormal rows/columns". For the case , matrices with orthonormal columns may be referred to as orthogonal k-frames and they are elements of the
Stiefel manifold In mathematics, the Stiefel manifold V_k(\R^n) is the set of all orthonormal ''k''-frames in \R^n. That is, it is the set of ordered orthonormal ''k''-tuples of vectors in \R^n. It is named after Swiss mathematician Eduard Stiefel. Likewise one ...
.


See also

*
Biorthogonal system In mathematics, a biorthogonal system is a pair of indexed families of vectors \tilde v_i \text E \text \tilde u_i \text F such that \left\langle\tilde v_i , \tilde u_j\right\rangle = \delta_, where E and F form a pair of topological vector spaces ...


Notes


References

* * * * *

* * *


External links

*
Tutorial and Interactive Program on Orthogonal Matrix
{{Matrix classes Matrices (mathematics)