Reducing Subspace
In linear algebra, a reducing subspace W of a linear map T:V\to V from a Hilbert space V to itself is an invariant subspace of T whose orthogonal complement W^\perp is also an invariant subspace of T. That is, T(W) \subseteq W and T(W^\perp) \subseteq W^\perp. One says that the subspace W reduces the map T. One says that a linear map is reducible if it has a nontrivial reducing subspace. Otherwise one says it is irreducible. If V is of finite dimension r and W is a reducing subspace of the map T:V\to V represented under basis B by matrix M \in\R^ then M can be expressed as the sum M = P_W M P_W + P_ M P_ where P_W \in\R^ is the matrix of the orthogonal projection from V to W and P_ = I - P_ is the matrix of the projection onto W^\perp. (Here I \in \R^ is the identity matrix.) Furthermore, V has an orthonormal basis B' with a subset that is an orthonormal basis of W. If Q \in \R^ is the transition matrix from B to B' then with respect to B' the matrix Q^MQ representing T is ... [...More Info...]       [...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]   |
|
Linear Algebra
Linear algebra is the branch of mathematics concerning linear equations such as: :a_1x_1+\cdots +a_nx_n=b, linear maps such as: :(x_1, \ldots, x_n) \mapsto a_1x_1+\cdots +a_nx_n, and their representations in vector spaces and through matrices. Linear algebra is central to almost all areas of mathematics. For instance, linear algebra is fundamental in modern presentations of geometry, including for defining basic objects such as lines, planes and rotations. Also, functional analysis, a branch of mathematical analysis, may be viewed as the application of linear algebra to spaces of functions. Linear algebra is also used in most sciences and fields of engineering, because it allows modeling many natural phenomena, and computing efficiently with such models. For nonlinear systems, which cannot be modeled with linear algebra, it is often used for dealing with first-order approximations, using the fact that the differential of a multivariate function at a point is the linear ma ... [...More Info...]       [...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]   |
|
Linear Map
In mathematics, and more specifically in linear algebra, a linear map (also called a linear mapping, linear transformation, vector space homomorphism, or in some contexts linear function) is a Map (mathematics), mapping V \to W between two vector spaces that preserves the operations of vector addition and scalar multiplication. The same names and the same definition are also used for the more general case of module (mathematics), modules over a ring (mathematics), ring; see Module homomorphism. If a linear map is a bijection then it is called a . In the case where V = W, a linear map is called a (linear) ''endomorphism''. Sometimes the term refers to this case, but the term "linear operator" can have different meanings for different conventions: for example, it can be used to emphasize that V and W are Real number, real vector spaces (not necessarily with V = W), or it can be used to emphasize that V is a function space, which is a common convention in functional analysis. Some ... [...More Info...]       [...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]   |
|
Hilbert Space
In mathematics, Hilbert spaces (named after David Hilbert) allow generalizing the methods of linear algebra and calculus from (finite-dimensional) Euclidean vector spaces to spaces that may be infinite-dimensional. Hilbert spaces arise naturally and frequently in mathematics and physics, typically as function spaces. Formally, a Hilbert space is a vector space equipped with an inner product that defines a distance function for which the space is a complete metric space. The earliest Hilbert spaces were studied from this point of view in the first decade of the 20th century by David Hilbert, Erhard Schmidt, and Frigyes Riesz. They are indispensable tools in the theories of partial differential equations, quantum mechanics, Fourier analysis (which includes applications to signal processing and heat transfer), and ergodic theory (which forms the mathematical underpinning of thermodynamics). John von Neumann coined the term ''Hilbert space'' for the abstract concept that under ... [...More Info...]       [...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]   |
|
Invariant Subspace
In mathematics, an invariant subspace of a linear mapping ''T'' : ''V'' → ''V '' i.e. from some vector space ''V'' to itself, is a subspace ''W'' of ''V'' that is preserved by ''T''; that is, ''T''(''W'') ⊆ ''W''. General description Consider a linear mapping T :T: W \to W. An invariant subspace W of T has the property that all vectors \mathbf \in W are transformed by T into vectors also contained in W. This can be stated as :\mathbf \in W \implies T(\mathbf) \in W. Trivial examples of invariant subspaces * \mathbb^n: Since T maps every vector in \mathbb^n into \mathbb^n. * \: Since a linear map has to map 0 \mapsto 0. 1-dimensional invariant subspace ''U'' A basis of a 1-dimensional space is simply a non-zero vector \mathbf. Consequently, any vector \mathbf \in U can be represented as \lambda \mathbf where \lambda is a scalar. If we represent T by a matrix A then, for U to be an invariant subspace it must satisfy : \forall \mathbf \in U \; \exists \alpha \in ... [...More Info...]       [...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]   |
|
Orthogonal Complement
In the mathematical fields of linear algebra and functional analysis, the orthogonal complement of a subspace ''W'' of a vector space ''V'' equipped with a bilinear form ''B'' is the set ''W''⊥ of all vectors in ''V'' that are orthogonal to every vector in ''W''. Informally, it is called the perp, short for perpendicular complement. It is a subspace of ''V''. Example Let V = (\R^5, \langle \cdot, \cdot \rangle) be the vector space equipped with the usual dot product \langle \cdot, \cdot \rangle (thus making it an inner product space), and let W = \, with A = \begin 1 & 0\\ 0 & 1\\ 2 & 6\\ 3 & 9\\ 5 & 3\\ \end. then its orthogonal complement W^\perp = \ can also be defined as W^\perp = \, being \tilde = \begin -2 & -3 & -5 \\ -6 & -9 & -3 \\ 1 & 0 & 0 \\ 0 & 1 & 0 \\ 0 & 0 & 1 \end. The fact that every column vector in A is orthogonal to every column vector in \tilde can be checked by direct computation. The fact that the spans of these vectors are orthogonal then follows by b ... [...More Info...]       [...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]   |
|
Linear Subspace
In mathematics, and more specifically in linear algebra, a linear subspace, also known as a vector subspaceThe term ''linear subspace'' is sometimes used for referring to flats and affine subspaces. In the case of vector spaces over the reals, linear subspaces, flats, and affine subspaces are also called ''linear manifolds'' for emphasizing that there are also manifolds. is a vector space that is a subset of some larger vector space. A linear subspace is usually simply called a ''subspace'' when the context serves to distinguish it from other types of subspaces. Definition If ''V'' is a vector space over a field ''K'' and if ''W'' is a subset of ''V'', then ''W'' is a linear subspace of ''V'' if under the operations of ''V'', ''W'' is a vector space over ''K''. Equivalently, a nonempty subset ''W'' is a subspace of ''V'' if, whenever are elements of ''W'' and are elements of ''K'', it follows that is in ''W''. As a corollary, all vector spaces are equipped with at least two ( ... [...More Info...]       [...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]   |
|
Dimension (vector Space)
In mathematics, the dimension of a vector space ''V'' is the cardinality (i.e., the number of vectors) of a basis of ''V'' over its base field. p. 44, §2.36 It is sometimes called Hamel dimension (after Georg Hamel) or algebraic dimension to distinguish it from other types of dimension. For every vector space there exists a basis, and all bases of a vector space have equal cardinality; as a result, the dimension of a vector space is uniquely defined. We say V is if the dimension of V is finite, and if its dimension is infinite. The dimension of the vector space V over the field F can be written as \dim_F(V) or as : F read "dimension of V over F". When F can be inferred from context, \dim(V) is typically written. Examples The vector space \R^3 has \left\ as a standard basis, and therefore \dim_(\R^3) = 3. More generally, \dim_(\R^n) = n, and even more generally, \dim_(F^n) = n for any field F. The complex numbers \Complex are both a real and complex vector space; we have ... [...More Info...]       [...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]   |
|
Basis (linear Algebra)
In mathematics, a set of vectors in a vector space is called a basis if every element of may be written in a unique way as a finite linear combination of elements of . The coefficients of this linear combination are referred to as components or coordinates of the vector with respect to . The elements of a basis are called . Equivalently, a set is a basis if its elements are linearly independent and every element of is a linear combination of elements of . In other words, a basis is a linearly independent spanning set. A vector space can have several bases; however all the bases have the same number of elements, called the ''dimension'' of the vector space. This article deals mainly with finite-dimensional vector spaces. However, many of the principles are also valid for infinite-dimensional vector spaces. Definition A basis of a vector space over a field (such as the real numbers or the complex numbers ) is a linearly independent subset of that spans . This me ... [...More Info...]       [...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]   |
|
Orthogonal Projection
In linear algebra and functional analysis, a projection is a linear transformation P from a vector space to itself (an endomorphism) such that P\circ P=P. That is, whenever P is applied twice to any vector, it gives the same result as if it were applied once (i.e. P is idempotent). It leaves its image unchanged. This definition of "projection" formalizes and generalizes the idea of graphical projection. One can also consider the effect of a projection on a geometrical object by examining the effect of the projection on points in the object. Definitions A projection on a vector space V is a linear operator P : V \to V such that P^2 = P. When V has an inner product and is complete (i.e. when V is a Hilbert space) the concept of orthogonality can be used. A projection P on a Hilbert space V is called an orthogonal projection if it satisfies \langle P \mathbf x, \mathbf y \rangle = \langle \mathbf x, P \mathbf y \rangle for all \mathbf x, \mathbf y \in V. A projection on a Hilber ... [...More Info...]       [...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]   |
|
Identity Matrix
In linear algebra, the identity matrix of size n is the n\times n square matrix with ones on the main diagonal and zeros elsewhere. Terminology and notation The identity matrix is often denoted by I_n, or simply by I if the size is immaterial or can be trivially determined by the context. I_1 = \begin 1 \end ,\ I_2 = \begin 1 & 0 \\ 0 & 1 \end ,\ I_3 = \begin 1 & 0 & 0 \\ 0 & 1 & 0 \\ 0 & 0 & 1 \end ,\ \dots ,\ I_n = \begin 1 & 0 & 0 & \cdots & 0 \\ 0 & 1 & 0 & \cdots & 0 \\ 0 & 0 & 1 & \cdots & 0 \\ \vdots & \vdots & \vdots & \ddots & \vdots \\ 0 & 0 & 0 & \cdots & 1 \end. The term unit matrix has also been widely used, but the term ''identity matrix'' is now standard. The term ''unit matrix'' is ambiguous, because it is also used for a matrix of ones and for any unit of the ring of all n\times n matrices. In some fields, such as group theory or quantum mechanics, the identity matrix is sometimes denoted by a boldface one, \mathbf, or called "id" (short for identity). ... [...More Info...]       [...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]   |
|
Orthonormal Basis
In mathematics, particularly linear algebra, an orthonormal basis for an inner product space ''V'' with finite dimension is a basis for V whose vectors are orthonormal, that is, they are all unit vectors and orthogonal to each other. For example, the standard basis for a Euclidean space \R^n is an orthonormal basis, where the relevant inner product is the dot product of vectors. The image of the standard basis under a rotation or reflection (or any orthogonal transformation) is also orthonormal, and every orthonormal basis for \R^n arises in this fashion. For a general inner product space V, an orthonormal basis can be used to define normalized orthogonal coordinates on V. Under these coordinates, the inner product becomes a dot product of vectors. Thus the presence of an orthonormal basis reduces the study of a finite-dimensional inner product space to the study of \R^n under dot product. Every finite-dimensional inner product space has an orthonormal basis, which may be ob ... [...More Info...]       [...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]   |
|
Change Of Basis
In mathematics, an ordered basis of a vector space of finite dimension allows representing uniquely any element of the vector space by a coordinate vector, which is a sequence of scalars called coordinates. If two different bases are considered, the coordinate vector that represents a vector on one basis is, in general, different from the coordinate vector that represents on the other basis. A change of basis consists of converting every assertion expressed in terms of coordinates relative to one basis into an assertion expressed in terms of coordinates relative to the other basis. Such a conversion results from the ''change-of-basis formula'' which expresses the coordinates relative to one basis in terms of coordinates relative to the other basis. Using matrices, this formula can be written :\mathbf x_\mathrm = A \,\mathbf x_\mathrm, where "old" and "new" refer respectively to the firstly defined basis and the other basis, \mathbf x_\mathrm and \mathbf x_\mathrm are the colu ... [...More Info...]       [...Related Items...]     OR:     [Wikipedia]   [Google]   [Baidu]   |