HOME

TheInfoList



OR:

An infinitesimal rotation matrix or differential rotation matrix is a
matrix Matrix (: matrices or matrixes) or MATRIX may refer to: Science and mathematics * Matrix (mathematics), a rectangular array of numbers, symbols or expressions * Matrix (logic), part of a formula in prenex normal form * Matrix (biology), the m ...
representing an
infinitely Infinity is something which is boundless, endless, or larger than any natural number. It is denoted by \infty, called the infinity symbol. From the time of the ancient Greeks, the philosophical nature of infinity has been the subject of man ...
small
rotation Rotation or rotational/rotary motion is the circular movement of an object around a central line, known as an ''axis of rotation''. A plane figure can rotate in either a clockwise or counterclockwise sense around a perpendicular axis intersect ...
. While a
rotation matrix In linear algebra, a rotation matrix is a transformation matrix that is used to perform a rotation (mathematics), rotation in Euclidean space. For example, using the convention below, the matrix :R = \begin \cos \theta & -\sin \theta \\ \sin \t ...
is an
orthogonal matrix In linear algebra, an orthogonal matrix, or orthonormal matrix, is a real square matrix whose columns and rows are orthonormal vectors. One way to express this is Q^\mathrm Q = Q Q^\mathrm = I, where is the transpose of and is the identi ...
R^\mathsf = R^ representing an element of SO(n) (the
special orthogonal group In mathematics, the orthogonal group in dimension , denoted , is the group of distance-preserving transformations of a Euclidean space of dimension that preserve a fixed point, where the group operation is given by composing transformations. ...
), the differential of a rotation is a
skew-symmetric matrix In mathematics, particularly in linear algebra, a skew-symmetric (or antisymmetric or antimetric) matrix is a square matrix whose transpose equals its negative. That is, it satisfies the condition In terms of the entries of the matrix, if a ...
A^\mathsf = -A in the
tangent space In mathematics, the tangent space of a manifold is a generalization of to curves in two-dimensional space and to surfaces in three-dimensional space in higher dimensions. In the context of physics the tangent space to a manifold at a point can be ...
\mathfrak(n) (the special orthogonal Lie algebra), which is not itself a rotation matrix. An infinitesimal rotation matrix has the form : I + d\theta \, A, where I is the identity matrix, d\theta is vanishingly small, and A \in \mathfrak(n). For example, if A = L_x, representing an infinitesimal three-dimensional rotation about the -axis, a basis element of \mathfrak(3), then : L_ = \begin 0 & 0 & 0 \\ 0 & 0 & -1 \\ 0 & 1 & 0 \end , and : I+d\theta L_ = \begin 1 & 0 & 0 \\ 0 & 1 & -d\theta \\ 0 & d\theta & 1 \end. The computation rules for infinitesimal rotation matrices are the usual ones except that infinitesimals of second order are dropped. With these rules, these matrices do not satisfy all the same properties as ordinary finite rotation matrices under the usual treatment of infinitesimals. It turns out that ''the order in which infinitesimal rotations are applied is irrelevant''.


Discussion

An infinitesimal rotation matrix is a
skew-symmetric matrix In mathematics, particularly in linear algebra, a skew-symmetric (or antisymmetric or antimetric) matrix is a square matrix whose transpose equals its negative. That is, it satisfies the condition In terms of the entries of the matrix, if a ...
where: * As any
rotation matrix In linear algebra, a rotation matrix is a transformation matrix that is used to perform a rotation (mathematics), rotation in Euclidean space. For example, using the convention below, the matrix :R = \begin \cos \theta & -\sin \theta \\ \sin \t ...
has a single real eigenvalue, which is equal to +1, the corresponding
eigenvector In linear algebra, an eigenvector ( ) or characteristic vector is a vector that has its direction unchanged (or reversed) by a given linear transformation. More precisely, an eigenvector \mathbf v of a linear transformation T is scaled by ...
defines the
rotation axis Rotation or rotational/rotary motion is the circular movement of an object around a central line, known as an ''axis of rotation''. A plane figure can rotate in either a clockwise or counterclockwise sense around a perpendicular axis intersect ...
. * Its module defines an infinitesimal
angular displacement The angular displacement (symbol θ, , or φ) – also called angle of rotation, rotational displacement, or rotary displacement – of a physical body is the angle (in units of radians, degrees, turns, etc.) through which the body rotates ( ...
. The shape of the matrix is as follows: A = \begin 1 & -d\phi_z(t) & d\phi_y(t) \\ d\phi_z(t) & 1 & -d\phi_x(t) \\ -d\phi_y(t) & d\phi_x(t) & 1 \\ \end


Associated quantities

Associated to an infinitesimal rotation matrix A is an '' infinitesimal rotation tensor'' d\Phi(t) = A - I: d\Phi(t) = \begin 0 & -d\phi_z(t) & d\phi_y(t) \\ d\phi_z(t) & 0 & -d\phi_x(t) \\ -d\phi_y(t) & d\phi_x(t) & 0 \\ \end Dividing it by the time difference yields the '' angular velocity tensor'': : \Omega = \frac = \begin 0 & -\omega_z(t) & \omega_y(t) \\ \omega_z(t) & 0 & -\omega_x(t) \\ -\omega_y(t) & \omega_x(t) & 0 \\ \end


Order of rotations

These matrices do not satisfy all the same properties as ordinary finite rotation matrices under the usual treatment of infinitesimals. To understand what this means, consider : dA_ = \begin 1 & 0 & 0 \\ 0 & 1 & -d\theta \\ 0 & d\theta & 1 \end. First, test the orthogonality condition, . The product is : dA_\mathbf^\textsf \, dA_\mathbf = \begin 1 & 0 & 0 \\ 0 & 1 + d\theta^2 & 0 \\ 0 & 0 & 1 + d\theta^2 \end, differing from an identity matrix by second-order infinitesimals, discarded here. So, to first order, an infinitesimal rotation matrix is an orthogonal matrix. Next, examine the square of the matrix, : dA_^2 = \begin 1 & 0 & 0 \\ 0 & 1 - d\theta^2 & -2d\theta \\ 0 & 2\,d\theta & 1 - d\theta^2 \end. Again discarding second-order effects, note that the angle simply doubles. This hints at the most essential difference in behavior, which we can exhibit with the assistance of a second infinitesimal rotation, :dA_\mathbf = \begin 1 & 0 & d\phi \\ 0 & 1 & 0 \\ -d\phi & 0 & 1 \end. Compare the products to , :\begin dA_\,dA_ &= \begin 1 & 0 & d\phi \\ d\theta\,d\phi & 1 & -d\theta \\ -d\phi & d\theta & 1 \end \\ dA_\,dA_ &= \begin 1 & d\theta\,d\phi & d\phi \\ 0 & 1 & -d\theta \\ -d\phi & d\theta & 1 \end. \\ \end Since d\theta \, d\phi is second-order, we discard it: thus, to first order, multiplication of infinitesimal rotation matrices is ''commutative''. In fact, : dA_\,dA_ = dA_\,dA_, again to first order. In other words, . This useful fact makes, for example, derivation of rigid body rotation relatively simple. But one must always be careful to distinguish (the first-order treatment of) these infinitesimal rotation matrices from both finite rotation matrices and from Lie algebra elements. When contrasting the behavior of finite rotation matrices in the
Baker–Campbell–Hausdorff formula In mathematics, the Baker–Campbell–Hausdorff formula gives the value of Z that solves the equation e^X e^Y = e^Z for possibly noncommutative and in the Lie algebra of a Lie group. There are various ways of writing the formula, but all ultima ...
above with that of infinitesimal rotation matrices, where all the commutator terms will be second-order infinitesimals, one finds a bona fide vector space. Technically, this dismissal of any second-order terms amounts to Group contraction.


Generators of rotations

Suppose we specify an axis of rotation by a unit vector 'x'', ''y'', ''z'' and suppose we have an infinitely small rotation of angle Δ''θ'' about that vector. Expanding the rotation matrix as an infinite addition, and taking the first-order approach, the rotation matrix Δ''R'' is represented as: : \Delta R = \begin 1 & 0 & 0 \\ 0 & 1 & 0 \\ 0 & 0 & 1 \end + \begin 0 & z & -y \\ -z & 0 & x \\ y & -x & 0 \end\,\Delta \theta = I + A\,\Delta\theta. A finite rotation through angle ''θ'' about this axis may be seen as a succession of small rotations about the same axis. Approximating Δ''θ'' as ''θ''/''N'', where ''N'' is a large number, a rotation of ''θ'' about the axis may be represented as: : R =\lim_ \left(I + \frac\right)^N = e^. It can be seen that Euler's theorem essentially states that ''all'' rotations may be represented in this form. The product ''Aθ'' is the "generator" of the particular rotation, being the vector associated with the matrix ''A''. This shows that the rotation matrix and the axis-angle format are related by the exponential function. One can derive a simple expression for the generator ''G''. One starts with an arbitrary plane defined by a pair of perpendicular unit vectors ''a'' and ''b''. In this plane one can choose an arbitrary vector ''x'' with perpendicular ''y''. One then solves for ''y'' in terms of ''x'' and substituting into an expression for a rotation in a plane yields the rotation matrix ''R'', which includes the generator . :\begin x &= a \cos\left( \alpha \right) + b \sin\left( \alpha \right) \\ y &= -a \sin\left( \alpha \right) + b \cos\left( \alpha \right) \\ \cos\left( \alpha \right) &= a^\mathrm x \\ \sin\left( \alpha \right) &= b^\mathrm x \\ y &= -ab^\mathrm x + ba^\mathrm x = \left( ba^\mathrm - ab^\mathrm \right)x \\ \\ x' &= x \cos\left( \beta \right) + y \sin\left( \beta \right) \\ &= \left I \cos\left( \beta \right) + \left( ba^\mathrm - ab^\mathrm \right) \sin\left( \beta \right) \right \\ \\ R &= I \cos\left( \beta \right) + \left( ba^\mathrm - ab^\mathrm \right) \sin\left( \beta \right) \\ &= I \cos\left( \beta \right) + G \sin\left( \beta \right) \\ \\ G &= ba^\mathrm - ab^\mathrm \\ \end To include vectors outside the plane in the rotation one needs to modify the above expression for ''R'' by including two projection operators that partition the space. This modified rotation matrix can be rewritten as an exponential function. : \begin P_ &= -G^2 \\ R &= I - P_ + \left I \cos\left( \beta \right) + G \sin\left( \beta \right) \rightP_ = e^ \\ \end Analysis is often easier in terms of these generators, rather than the full rotation matrix. Analysis in terms of the generators is known as the
Lie algebra In mathematics, a Lie algebra (pronounced ) is a vector space \mathfrak g together with an operation called the Lie bracket, an alternating bilinear map \mathfrak g \times \mathfrak g \rightarrow \mathfrak g, that satisfies the Jacobi ident ...
of the rotation group.


Exponential map

Connecting the Lie algebra to the Lie group is the exponential map, which is defined using the standard
matrix exponential In mathematics, the matrix exponential is a matrix function on square matrix, square matrices analogous to the ordinary exponential function. It is used to solve systems of linear differential equations. In the theory of Lie groups, the matrix exp ...
series for For any
skew-symmetric matrix In mathematics, particularly in linear algebra, a skew-symmetric (or antisymmetric or antimetric) matrix is a square matrix whose transpose equals its negative. That is, it satisfies the condition In terms of the entries of the matrix, if a ...
, is always a rotation matrix. An important practical example is the case. In
rotation group SO(3) In mechanics and geometry, the 3D rotation group, often denoted SO(3), is the group of all rotations about the origin of three-dimensional Euclidean space \R^3 under the operation of composition. By definition, a rotation about the origin is a ...
, it is shown that one can identify every with an Euler vector , where is a unit magnitude vector. By the properties of the identification , is in the null space of . Thus, is left invariant by and is hence a rotation axis. Using Rodrigues' rotation formula on matrix form with , together with standard double angle formulae one obtains, : \begin \exp( A ) &= \exp(\theta(\boldsymbol)) = \exp \left( \left begin 0 & -z \theta & y \theta \\ z \theta & 0&-x \theta \\ -y \theta & x \theta & 0 \end\right\right)= \boldsymbol + 2\cos\frac\sin\frac~\boldsymbol + 2\sin^2\frac ~(\boldsymbol )^2 , \end This is the matrix for a rotation around axis by the angle in half-angle form. For full detail, see exponential map SO(3). Notice that for infinitesimal angles second-order terms can be ignored and remains


Relationship to skew-symmetric matrices

Skew-symmetric matrices over the field of real numbers form the
tangent space In mathematics, the tangent space of a manifold is a generalization of to curves in two-dimensional space and to surfaces in three-dimensional space in higher dimensions. In the context of physics the tangent space to a manifold at a point can be ...
to the real
orthogonal group In mathematics, the orthogonal group in dimension , denoted , is the Group (mathematics), group of isometry, distance-preserving transformations of a Euclidean space of dimension that preserve a fixed point, where the group operation is given by ...
O(n) at the identity matrix; formally, the special orthogonal Lie algebra. In this sense, then, skew-symmetric matrices can be thought of as ''infinitesimal rotations''. Another way of saying this is that the space of skew-symmetric matrices forms the
Lie algebra In mathematics, a Lie algebra (pronounced ) is a vector space \mathfrak g together with an operation called the Lie bracket, an alternating bilinear map \mathfrak g \times \mathfrak g \rightarrow \mathfrak g, that satisfies the Jacobi ident ...
o(n) of the
Lie group In mathematics, a Lie group (pronounced ) is a group (mathematics), group that is also a differentiable manifold, such that group multiplication and taking inverses are both differentiable. A manifold is a space that locally resembles Eucli ...
O(n). The Lie bracket on this space is given by the
commutator In mathematics, the commutator gives an indication of the extent to which a certain binary operation fails to be commutative. There are different definitions used in group theory and ring theory. Group theory The commutator of two elements, ...
: :
, B The comma is a punctuation mark that appears in several variants in different languages. Some typefaces render it as a small line, slightly curved or straight, but inclined from the vertical; others give it the appearance of a miniature fille ...
= AB - BA.\, It is easy to check that the commutator of two skew-symmetric matrices is again skew-symmetric: : \begin A, B^\textsf &= B^\textsf A^\textsf - A^\textsf B^\textsf \\ &= (-B)(-A) - (-A)(-B) = BA - AB = -
, B The comma is a punctuation mark that appears in several variants in different languages. Some typefaces render it as a small line, slightly curved or straight, but inclined from the vertical; others give it the appearance of a miniature fille ...
\, . \end The
matrix exponential In mathematics, the matrix exponential is a matrix function on square matrix, square matrices analogous to the ordinary exponential function. It is used to solve systems of linear differential equations. In the theory of Lie groups, the matrix exp ...
of a skew-symmetric matrix A is then an
orthogonal matrix In linear algebra, an orthogonal matrix, or orthonormal matrix, is a real square matrix whose columns and rows are orthonormal vectors. One way to express this is Q^\mathrm Q = Q Q^\mathrm = I, where is the transpose of and is the identi ...
R: :R = \exp(A) = \sum_^\infty \frac. The image of the exponential map of a Lie algebra always lies in the connected component of the Lie group that contains the identity element. In the case of the Lie group O(n), this connected component is the
special orthogonal group In mathematics, the orthogonal group in dimension , denoted , is the group of distance-preserving transformations of a Euclidean space of dimension that preserve a fixed point, where the group operation is given by composing transformations. ...
SO(n), consisting of all orthogonal matrices with determinant 1. So R = \exp(A) will have determinant +1. Moreover, since the exponential map of a connected compact Lie group is always surjective, it turns out that ''every'' orthogonal matrix with unit determinant can be written as the exponential of some skew-symmetric matrix. In the particular important case of dimension n=2, the exponential representation for an orthogonal matrix reduces to the well-known
polar form In mathematics, a complex number is an element of a number system that extends the real numbers with a specific element denoted , called the imaginary unit and satisfying the equation i^= -1; every complex number can be expressed in the form ...
of a complex number of unit modulus. Indeed, if n=2, a special orthogonal matrix has the form :\begin a & -b \\ b & \,a \end, with a^2 + b^2 = 1. Therefore, putting a = \cos\theta and b = \sin\theta, it can be written :\begin \cos\,\theta & -\sin\,\theta \\ \sin\,\theta & \,\cos\,\theta \end = \exp\left(\theta\begin 0 & -1 \\ 1 & \,0 \end\right), which corresponds exactly to the polar form \cos \theta + i \sin \theta = e^ of a complex number of unit modulus. In 3 dimensions, the matrix exponential is Rodrigues' rotation formula in matrix notation, and when expressed via the Euler-Rodrigues formula, the algebra of its four parameters gives rise to quaternions. The exponential representation of an orthogonal matrix of order n can also be obtained starting from the fact that in dimension n any special orthogonal matrix R can be written as R = QSQ^\textsf, where Q is orthogonal and S is a
block diagonal matrix In mathematics, a block matrix or a partitioned matrix is a matrix that is interpreted as having been broken into sections called blocks or submatrices. Intuitively, a matrix interpreted as a block matrix can be visualized as the original matrix w ...
with \lfloor n/2\rfloor blocks of order 2, plus one of order 1 if n is odd; since each single block of order 2 is also an orthogonal matrix, it admits an exponential form. Correspondingly, the matrix ''S'' writes as exponential of a skew-symmetric block matrix \Sigma of the form above, S = \exp(\Sigma), so that R = Q\exp(\Sigma)Q^\textsf = \exp(Q\Sigma Q^\textsf), exponential of the skew-symmetric matrix Q\Sigma Q^\textsf. Conversely, the surjectivity of the exponential map, together with the above-mentioned block-diagonalization for skew-symmetric matrices, implies the block-diagonalization for orthogonal matrices.


See also

* Generators of rotations * Infinitesimal rotations * Infinitesimal rotation tensor *
Infinitesimal transformation In mathematics, an infinitesimal transformation is a limiting form of ''small'' transformation. For example one may talk about an infinitesimal rotation of a rigid body, in three-dimensional space. This is conventionally represented by a 3×3 ...
* Rotation group SO(3)#Infinitesimal rotations


Notes


References


Sources

* * {{Citation , last=Wedderburn , first=Joseph H. M. , author-link=Joseph Wedderburn , year=1934 , title=Lectures on Matrices , publisher=
AMS AMS or Ams may refer to: Organizations Companies * Alenia Marconi Systems * American Management Systems * AMS (Advanced Music Systems) * ams AG, semiconductor manufacturer * AMS Pictures * Auxiliary Medical Services Educational institutions ...
, isbn=978-0-8218-3204-2 , url=https://scholar.google.co.uk/scholar?hl=en&lr=&q=author%3AWedderburn+intitle%3ALectures+on+Matrices&as_publication=&as_ylo=1934&as_yhi=1934&btnG=Search Rotation Mathematics of infinitesimals