HOME

TheInfoList



OR:

The rank–nullity theorem is a theorem in
linear algebra Linear algebra is the branch of mathematics concerning linear equations such as :a_1x_1+\cdots +a_nx_n=b, linear maps such as :(x_1, \ldots, x_n) \mapsto a_1x_1+\cdots +a_nx_n, and their representations in vector spaces and through matrix (mathemat ...
, which asserts: * the number of columns of a matrix is the sum of the rank of and the nullity of ; and * the
dimension In physics and mathematics, the dimension of a mathematical space (or object) is informally defined as the minimum number of coordinates needed to specify any point within it. Thus, a line has a dimension of one (1D) because only one coo ...
of the domain of a
linear transformation In mathematics, and more specifically in linear algebra, a linear map (also called a linear mapping, linear transformation, vector space homomorphism, or in some contexts linear function) is a mapping V \to W between two vector spaces that pr ...
is the sum of the rank of (the dimension of the
image An image or picture is a visual representation. An image can be Two-dimensional space, two-dimensional, such as a drawing, painting, or photograph, or Three-dimensional space, three-dimensional, such as a carving or sculpture. Images may be di ...
of ) and the nullity of (the dimension of the kernel of ). p. 70, §2.1, Theorem 2.3 It follows that for linear transformations of
vector space In mathematics and physics, a vector space (also called a linear space) is a set (mathematics), set whose elements, often called vector (mathematics and physics), ''vectors'', can be added together and multiplied ("scaled") by numbers called sc ...
s of equal finite dimension, either
injectivity In mathematics, an injective function (also known as injection, or one-to-one function ) is a function that maps distinct elements of its domain to distinct elements of its codomain; that is, implies (equivalently by contraposition, impl ...
or surjectivity implies bijectivity.


Stating the theorem


Linear transformations

Let T : V \to W be a linear transformation between two vector spaces where T's domain V is finite dimensional. Then \operatorname(T) ~+~ \operatorname(T) ~=~ \dim V, where \operatorname(T) is the rank of T (the
dimension In physics and mathematics, the dimension of a mathematical space (or object) is informally defined as the minimum number of coordinates needed to specify any point within it. Thus, a line has a dimension of one (1D) because only one coo ...
of its
image An image or picture is a visual representation. An image can be Two-dimensional space, two-dimensional, such as a drawing, painting, or photograph, or Three-dimensional space, three-dimensional, such as a carving or sculpture. Images may be di ...
) and \operatorname(T) is the nullity of T (the dimension of its kernel). In other words, \dim (\operatorname T) + \dim (\operatorname T) = \dim (\operatorname(T)). This theorem can be refined via the splitting lemma to be a statement about an
isomorphism In mathematics, an isomorphism is a structure-preserving mapping or morphism between two structures of the same type that can be reversed by an inverse mapping. Two mathematical structures are isomorphic if an isomorphism exists between the ...
of spaces, not just dimensions. Explicitly, since T induces an isomorphism from V / \operatorname (T) to \operatorname (T), the existence of a basis for V that extends any given basis of \operatorname(T) implies, via the splitting lemma, that \operatorname(T) \oplus \operatorname(T) \cong V. Taking dimensions, the rank–nullity theorem follows.


Matrices

Linear maps can be represented with
matrices Matrix (: matrices or matrixes) or MATRIX may refer to: Science and mathematics * Matrix (mathematics), a rectangular array of numbers, symbols or expressions * Matrix (logic), part of a formula in prenex normal form * Matrix (biology), the ...
. More precisely, an m\times n matrix represents a linear map f:F^n\to F^m, where F is the underlying field. So, the dimension of the domain of f is , the number of columns of , and the rank–nullity theorem for an m\times n matrix is \operatorname(M) + \operatorname(M) = n.


Proofs

Here we provide two proofs. The first operates in the general case, using linear maps. The second proof looks at the homogeneous system \mathbf = \mathbf, where \mathbf is a m\times n with rank r, and shows explicitly that there exists a set of n-r
linearly independent In the theory of vector spaces, a set of vectors is said to be if there exists no nontrivial linear combination of the vectors that equals the zero vector. If such a linear combination exists, then the vectors are said to be . These concep ...
solutions that span the null space of \mathbf. While the theorem requires that the domain of the linear map be finite-dimensional, there is no such assumption on the codomain. This means that there are linear maps not given by matrices for which the theorem applies. Despite this, the first proof is not actually more general than the second: since the image of the linear map is finite-dimensional, we can represent the map from its domain to its image by a matrix, prove the theorem for that matrix, then compose with the inclusion of the image into the full codomain.


First proof

Let V, W be vector spaces over some field F, and T defined as in the statement of the theorem with \dim V = n. As \operatornameT \subset V is a subspace, there exists a basis for it. Suppose \dim\operatornameT = k and let \mathcal := \ \subset \operatorname(T) be such a basis. We may now, by the Steinitz exchange lemma, extend \mathcal with n-k linearly independent vectors w_1, \ldots, w_ to form a full basis of V. Let \mathcal := \ \subset V \setminus \operatorname(T) such that \mathcal := \mathcal \cup \mathcal = \ \subset V is a basis for V. From this, we know that \operatorname T = \operatornameT(\mathcal) = \operatorname\ :: = \operatorname\ = \operatornameT(\mathcal) . We now claim that T(\mathcal) is a basis for \operatorname T. The above equality already states that T(\mathcal) is a generating set for \operatorname T; it remains to be shown that it is also linearly independent to conclude that it is a basis. Suppose T(\mathcal) is not linearly independent, and let \sum_^ \alpha _j T(w_j) = 0_W for some \alpha _j \in F. Thus, owing to the linearity of T, it follows that T \left(\sum_^ \alpha _j w_j \right) = 0_W \implies \left(\sum_^ \alpha _j w_j \right) \in \operatorname T = \operatorname \mathcal \subset V . This is a contradiction to \mathcal being a basis, unless all \alpha _j are equal to zero. This shows that T(\mathcal) is linearly independent, and more specifically that it is a basis for \operatornameT. To summarize, we have \mathcal, a basis for \operatornameT, and T(\mathcal), a basis for \operatornameT. Finally we may state that \operatorname(T) + \operatorname(T) = \dim \operatorname T + \dim \operatornameT :: = , T(\mathcal), + , \mathcal, = (n-k) + k = n = \dim V . This concludes our proof.


Second proof

Let \mathbf be an m\times n matrix with r
linearly independent In the theory of vector spaces, a set of vectors is said to be if there exists no nontrivial linear combination of the vectors that equals the zero vector. If such a linear combination exists, then the vectors are said to be . These concep ...
columns (i.e. \operatorname(\mathbf) = r). We will show that: To do this, we will produce an n \times (n-r) matrix \mathbf whose columns form a basis of the null space of \mathbf. Without loss of generality, assume that the first r columns of \mathbf are linearly independent. So, we can write \mathbf = \begin \mathbf_1 & \mathbf_2\end , where *\mathbf_1 is an m \times r matrix with r linearly independent column vectors, and *\mathbf_2 is an m\times (n-r) matrix such that each of its n-r columns is linear combinations of the columns of \mathbf_1. This means that \mathbf_2 = \mathbf_1\mathbf for some r \times (n-r) matrix \mathbf (see
rank factorization A rank is a position in a hierarchy. It can be formally recognized—for example, cardinal, chief executive officer, general, professor—or unofficial. People Formal ranks * Academic rank * Corporate title * Diplomatic rank * Hierarch ...
) and, hence, \mathbf = \begin \mathbf_1 & \mathbf_1\mathbf\end . Let \mathbf = \begin -\mathbf \\ \mathbf_ \end , where \mathbf_ is the (n-r)\times (n-r)
identity matrix In linear algebra, the identity matrix of size n is the n\times n square matrix with ones on the main diagonal and zeros elsewhere. It has unique properties, for example when the identity matrix represents a geometric transformation, the obje ...
. So, \mathbf is an n \times (n-r) matrix such that \mathbf\mathbf = \begin\mathbf_1 & \mathbf_1\mathbf \end\begin -\mathbf \\ \mathbf_ \end = -\mathbf_1\mathbf + \mathbf_1\mathbf = \mathbf_. Therefore, each of the n-r columns of \mathbf are particular solutions of \mathbf = _. Furthermore, the n-r columns of \mathbf are
linearly independent In the theory of vector spaces, a set of vectors is said to be if there exists no nontrivial linear combination of the vectors that equals the zero vector. If such a linear combination exists, then the vectors are said to be . These concep ...
because \mathbf = \mathbf_ will imply \mathbf = \mathbf_ for \mathbf \in ^: \mathbf\mathbf = \mathbf_ \implies \begin -\mathbf \\ \mathbf_ \end\mathbf = \mathbf_ \implies \begin -\mathbf\mathbf \\ \mathbf \end = \begin \mathbf_ \\ \mathbf_ \end \implies \mathbf = \mathbf_. Therefore, the column vectors of \mathbf constitute a set of n-r linearly independent solutions for \mathbf = \mathbf_. We next prove that ''any'' solution of \mathbf = \mathbf_ must be a
linear combination In mathematics, a linear combination or superposition is an Expression (mathematics), expression constructed from a Set (mathematics), set of terms by multiplying each term by a constant and adding the results (e.g. a linear combination of ''x'' a ...
of the columns of \mathbf. For this, let \mathbf = \begin \mathbf_1 \\ \mathbf_2 \end \in ^ be any vector such that \mathbf = \mathbf_. Since the columns of \mathbf_1 are linearly independent, \mathbf_1\mathbf = \mathbf_ implies \mathbf = \mathbf_. Therefore, \begin \mathbf\mathbf & = & \mathbf_ \\ \implies \begin\mathbf_1 & \mathbf_1\mathbf\end \begin \mathbf_1 \\ \mathbf_2 \end & = & \mathbf_1\mathbf_1 + \mathbf_1\mathbf\mathbf_2 & = & \mathbf_1(\mathbf_1 + \mathbf\mathbf_2) & = & \mathbf_ \\ \implies \mathbf_1 + \mathbf\mathbf_2 & = & \mathbf_ \\ \implies \mathbf_1 & = & -\mathbf\mathbf_2 \end \implies \mathbf = \begin \mathbf_1 \\ \mathbf_2 \end = \begin -\mathbf \\ \mathbf_ \end\mathbf_2 = \mathbf\mathbf_2. This proves that any vector \mathbf that is a solution of \mathbf = \mathbf must be a linear combination of the n-r special solutions given by the columns of \mathbf. And we have already seen that the columns of \mathbf are linearly independent. Hence, the columns of \mathbf constitute a basis for the null space of \mathbf. Therefore, the nullity of \mathbf is n - r. Since r equals rank of \mathbf, it follows that \operatorname(\mathbf) + \operatorname(\mathbf) = n. This concludes our proof.


A third fundamental subspace

When T: V \to W is a linear transformation between two finite-dimensional subspaces, with n = \dim(V) and m = \dim(W) (so can be represented by an m \times n matrix M), the rank–nullity theorem asserts that if T has rank r, then n - r is the dimension of the null space of M, which represents the kernel of T. In some texts, a third fundamental subspace associated to T is considered alongside its image and kernel: the
cokernel The cokernel of a linear mapping of vector spaces is the quotient space of the codomain of by the image of . The dimension of the cokernel is called the ''corank'' of . Cokernels are dual to the kernels of category theory, hence the nam ...
of T is the quotient space W / \operatorname(T), and its dimension is m - r. This dimension formula (which might also be rendered \dim \operatorname(T) + \dim\operatorname(T) = \dim(W)) together with the rank–nullity theorem is sometimes called the ''fundamental theorem of linear algebra''.


Reformulations and generalizations

This theorem is a statement of the
first isomorphism theorem In mathematics, specifically abstract algebra, the isomorphism theorems (also known as Noether's isomorphism theorems) are theorems that describe the relationship among quotients, homomorphisms, and subobjects. Versions of the theorems exist for ...
of algebra for the case of vector spaces; it generalizes to the splitting lemma. In more modern language, the theorem can also be phrased as saying that each short exact sequence of vector spaces splits. Explicitly, given that 0 \rightarrow U \rightarrow V \mathbin R \rightarrow 0 is a
short exact sequence In mathematics, an exact sequence is a sequence of morphisms between objects (for example, Group (mathematics), groups, Ring (mathematics), rings, Module (mathematics), modules, and, more generally, objects of an abelian category) such that the Im ...
of vector spaces, then U \oplus R \cong V , hence \dim(U) + \dim(R) = \dim(V) . Here R plays the role of \operatorname T and U is \operatornameT, i.e. 0 \rightarrow \ker T \mathbin V \mathbin \operatorname T \rightarrow 0 In the finite-dimensional case, this formulation is susceptible to a generalization: if 0 \rightarrow V_1 \rightarrow V_2 \rightarrow \cdots V_r \rightarrow 0 is an
exact sequence In mathematics, an exact sequence is a sequence of morphisms between objects (for example, groups, rings, modules, and, more generally, objects of an abelian category) such that the image of one morphism equals the kernel of the next. Definit ...
of finite-dimensional vector spaces, then \sum_^r (-1)^i\dim(V_i) = 0. The rank–nullity theorem for finite-dimensional vector spaces may also be formulated in terms of the ''index'' of a linear map. The index of a linear map T \in \operatorname(V,W), where V and W are finite-dimensional, is defined by \operatorname T = \dim \operatorname(T) - \dim \operatorname T . Intuitively, \dim \operatorname T is the number of independent solutions v of the equation Tv = 0, and \dim \operatorname T is the number of independent restrictions that have to be put on w to make Tv = w solvable. The rank–nullity theorem for finite-dimensional vector spaces is equivalent to the statement \operatorname T = \dim V - \dim W . We see that we can easily read off the index of the linear map T from the involved spaces, without any need to analyze T in detail. This effect also occurs in a much deeper result: the Atiyah–Singer index theorem states that the index of certain differential operators can be read off the geometry of the involved spaces.


Citations


References

* * * *. * *


External links

*
MIT Linear Algebra Lecture on the Four Fundamental Subspaces
from MIT OpenCourseWare {{DEFAULTSORT:Rank-nullity theorem Theorems in linear algebra Isomorphism theorems Articles containing proofs