HOME

TheInfoList



OR:

In
mathematics Mathematics is a field of study that discovers and organizes methods, Mathematical theory, theories and theorems that are developed and Mathematical proof, proved for the needs of empirical sciences and mathematics itself. There are many ar ...
, specifically
linear algebra Linear algebra is the branch of mathematics concerning linear equations such as :a_1x_1+\cdots +a_nx_n=b, linear maps such as :(x_1, \ldots, x_n) \mapsto a_1x_1+\cdots +a_nx_n, and their representations in vector spaces and through matrix (mathemat ...
, the Cauchy–Binet formula, named after
Augustin-Louis Cauchy Baron Augustin-Louis Cauchy ( , , ; ; 21 August 1789 – 23 May 1857) was a French mathematician, engineer, and physicist. He was one of the first to rigorously state and prove the key theorems of calculus (thereby creating real a ...
and Jacques Philippe Marie Binet, is an identity for the
determinant In mathematics, the determinant is a Scalar (mathematics), scalar-valued function (mathematics), function of the entries of a square matrix. The determinant of a matrix is commonly denoted , , or . Its value characterizes some properties of the ...
of the product of two rectangular
matrices Matrix (: matrices or matrixes) or MATRIX may refer to: Science and mathematics * Matrix (mathematics), a rectangular array of numbers, symbols or expressions * Matrix (logic), part of a formula in prenex normal form * Matrix (biology), the ...
of transpose shapes (so that the product is well-defined and
square In geometry, a square is a regular polygon, regular quadrilateral. It has four straight sides of equal length and four equal angles. Squares are special cases of rectangles, which have four equal angles, and of rhombuses, which have four equal si ...
). It generalizes the statement that the determinant of a product of square matrices is equal to the product of their determinants. The formula is valid for matrices with the entries from any
commutative ring In mathematics, a commutative ring is a Ring (mathematics), ring in which the multiplication operation is commutative. The study of commutative rings is called commutative algebra. Complementarily, noncommutative algebra is the study of ring prope ...
.


Statement

Let ''A'' be an ''m''×''n'' matrix and ''B'' an ''n''×''m'' matrix. Write 'n''for the set , and \tbinomm for the set of ''m''-
combination In mathematics, a combination is a selection of items from a set that has distinct members, such that the order of selection does not matter (unlike permutations). For example, given three fruits, say an apple, an orange and a pear, there are ...
s of 'n''(i.e., subsets of 'n''of size ''m''; there are \tbinom nm of them). For S\in\tbinomm, write ''A'' 'm''''S'' for the ''m''×''m'' matrix whose columns are the columns of ''A'' at indices from ''S'', and ''B''''S'', 'm''/sub> for the ''m''×''m'' matrix whose rows are the rows of ''B'' at indices from ''S''. The Cauchy–Binet formula then states : \det(AB) = \sum_ \det(A_)\det(B_). Example: Taking ''m'' = 2 and ''n'' = 3, and matrices A = \begin1&1&2\\ 3& 1& -1\\ \end and B = \begin1&1\\3&1\\0&2\end, the Cauchy–Binet formula gives the determinant : \det(AB)= \left, \begin1&1\\3&1\end\ \cdot \left, \begin1&1\\3&1\end\ + \left, \begin1&2\\1&-1\end\ \cdot \left, \begin3&1\\0&2\end\ + \left, \begin1&2\\3&-1\end\ \cdot \left, \begin1&1\\0&2\end\. Indeed AB =\begin4&6\\6&2\end, and its determinant is -28 which equals (-2) \times (-2) + (-3) \times 6 + (-7) \times 2 from the right hand side of the formula.


Special cases

If ''n'' < ''m'' then \tbinomm is the empty set, and the formula says that det(''AB'') = 0 (its right hand side is an
empty sum In mathematics, an empty sum, or nullary sum, is a summation where the number of terms is zero. The natural way to extend non-empty sums is to let the empty sum be the additive identity. Let a_1, a_2, a_3, ... be a sequence of numbers, and let ...
); indeed in this case the rank of the ''m''×''m'' matrix ''AB'' is at most ''n'', which implies that its determinant is zero. If ''n'' = ''m'', the case where ''A'' and ''B'' are square matrices, \tbinomm=\ (a singleton set), so the sum only involves ''S'' =  'n'' and the formula states that det(''AB'') = det(''A'')det(''B''). For ''m'' = 0, ''A'' and ''B'' are empty matrices (but of different shapes if ''n'' > 0), as is their product ''AB''; the summation involves a single term ''S'' = Ø, and the formula states 1 = 1, with both sides given by the determinant of the 0×0 matrix. For ''m'' = 1, the summation ranges over the collection \tbinom1 of the ''n'' different singletons taken from 'n'' and both sides of the formula give \textstyle\sum_^nA_B_, the
dot product In mathematics, the dot product or scalar productThe term ''scalar product'' means literally "product with a Scalar (mathematics), scalar as a result". It is also used for other symmetric bilinear forms, for example in a pseudo-Euclidean space. N ...
of the pair of
vector Vector most often refers to: * Euclidean vector, a quantity with a magnitude and a direction * Disease vector, an agent that carries and transmits an infectious pathogen into another living organism Vector may also refer to: Mathematics a ...
s represented by the matrices. The smallest value of ''m'' for which the formula states a non-trivial equality is ''m'' = 2; it is discussed in the article on the Binet–Cauchy identity.


In the case ''n'' = 3

Let \boldsymbol, \boldsymbol, \boldsymbol, \boldsymbol, \boldsymbol, \boldsymbol,\boldsymbol,\boldsymbol be three-dimensional vectors. : \begin & 1 = 1 & (m = 0)\\ 0pt& \boldsymbol\cdot \boldsymbol = a_1 x_1 + a_2 x_2 + a_3 x_3 & (m = 1)\\ 0pt& \begin \boldsymbol\cdot \boldsymbol & \boldsymbol\cdot \boldsymbol\\ \boldsymbol\cdot \boldsymbol & \boldsymbol\cdot \boldsymbol \end \\ pt= & \begin a_2 & a_3\\ b_2 & b_3 \end \begin x_2 & y_2\\ x_3 & y_3 \end + \begin a_3 & a_1\\ b_3 & b_1 \end \begin x_3 & y_3\\ x_1 & y_1 \end + \begin a_1 & a_2\\ b_1 & b_2 \end \begin x_1 & y_1\\ x_2 & y_2 \end\\ pt= & (\boldsymbol\times\boldsymbol)\cdot(\boldsymbol\times\boldsymbol) & (m = 2)\\ 0pt& \begin \boldsymbol\cdot \boldsymbol & \boldsymbol\cdot \boldsymbol & \boldsymbol\cdot \boldsymbol\\ \boldsymbol\cdot \boldsymbol & \boldsymbol\cdot \boldsymbol & \boldsymbol\cdot \boldsymbol\\ \boldsymbol\cdot \boldsymbol & \boldsymbol\cdot \boldsymbol & \boldsymbol\cdot \boldsymbol \end = \begin a_1 & a_2 & a_3\\ b_1 & b_2 & b_3\\ c_1 & c_2 & c_3 \end \begin x_1 & y_1 & z_1 \\ x_2 & y_2 & z_2 \\ x_3 & y_3 & z_3 \end\\ pt= & boldsymbol\cdot (\boldsymbol \times \boldsymbol) boldsymbol\cdot (\boldsymbol \times \boldsymbol) & (m = 3)\\ 0pt& \begin \boldsymbol\cdot \boldsymbol & \boldsymbol\cdot \boldsymbol & \boldsymbol\cdot \boldsymbol & \boldsymbol\cdot \boldsymbol \\ \boldsymbol\cdot \boldsymbol & \boldsymbol\cdot \boldsymbol & \boldsymbol\cdot \boldsymbol & \boldsymbol\cdot \boldsymbol \\ \boldsymbol\cdot \boldsymbol & \boldsymbol\cdot \boldsymbol & \boldsymbol\cdot \boldsymbol & \boldsymbol\cdot \boldsymbol \\ \boldsymbol\cdot \boldsymbol & \boldsymbol\cdot \boldsymbol & \boldsymbol\cdot \boldsymbol & \boldsymbol\cdot \boldsymbol \end = 0 & (m = 4) \end In the case ''m'' > 3, the right-hand side always equals 0.


A simple proof

The following simple proof relies on two facts that can be proven in several different ways: # For any 1 \leq k\leq n the coefficient of z^ in the polynomial \det(zI_n+X) is the sum of the k\times k principal minors of X. # If m \leq n and A is an m\times n matrix and B an n\times m matrix, then ::: \det(zI_n+BA)=z^\det(zI_m+AB). Now, if we compare the coefficient of z^ in the equation \det(zI_n+BA)=z^\det(zI_m+AB), the left hand side will give the sum of the principal minors of BA while the right hand side will give the constant term of \det(zI_m+AB), which is simply \det(AB), which is what the Cauchy–Binet formula states, i.e. : \begin &\det(AB)= \sum_ \det((BA)_)=\sum_ \det(B_) \det(A_) \\ pt= & \sum_\det(A_) \det(B_). \end


Proof

There are various kinds of proofs that can be given for the Cauchy−Binet formula. The proof below is based on formal manipulations only, and avoids using any particular interpretation of determinants, which may be taken to be defined by the Leibniz formula. Only their multilinearity with respect to rows and columns, and their alternating property (vanishing in the presence of equal rows or columns) are used; in particular the multiplicative property of determinants for square matrices is not used, but is rather established (the case ''n'' = ''m''). The proof is valid for arbitrary commutative coefficient rings. The formula can be proved in two steps: # use the fact that both sides are multilinear (more precisely 2''m''-linear) in the ''rows'' of ''A'' and the ''columns'' of ''B'', to reduce to the case that each row of ''A'' and each column of ''B'' has only one non-zero entry, which is 1. # handle that case using the functions 'm''nbsp;→  'n''that map respectively the row numbers of ''A'' to the column number of their nonzero entry, and the column numbers of ''B'' to the row number of their nonzero entry. For step 1, observe that for each row of ''A'' or column of ''B'', and for each ''m''-combination ''S'', the values of det(''AB'') and det(''A'' 'm''''S'')det(''B''''S'', 'm''/sub>) indeed depend linearly on the row or column. For the latter this is immediate from the multilinear property of the determinant; for the former one must in addition check that taking a linear combination for the row of ''A'' or column of ''B'' while leaving the rest unchanged only affects the corresponding row or column of the product ''AB'', and by the same linear combination. Thus one can work out both sides of the Cauchy−Binet formula by linearity for every row of ''A'' and then also every column of ''B'', writing each of the rows and columns as a linear combination of standard basis vectors. The resulting multiple summations are huge, but they have the same form for both sides: corresponding terms involve the same scalar factor (each is a product of entries of ''A'' and of ''B''), and these terms only differ by involving two different expressions in terms of constant matrices of the kind described above, which expressions should be equal according to the Cauchy−Binet formula. This achieves the reduction of the first step. Concretely, the multiple summations can be grouped into two summations, one over all functions ''f'': 'm''nbsp;→  'n''that for each row index of ''A'' gives a corresponding column index, and one over all functions ''g'': 'm''nbsp;→  'n''that for each column index of ''B'' gives a corresponding row index. The matrices associated to ''f'' and ''g'' are :L_f=\bigl((\delta_)_\bigr) \quad\text \quad R_g=\bigl((\delta_)_\bigr) where "\delta" is the
Kronecker delta In mathematics, the Kronecker delta (named after Leopold Kronecker) is a function of two variables, usually just non-negative integers. The function is 1 if the variables are equal, and 0 otherwise: \delta_ = \begin 0 &\text i \neq j, \\ 1 &\ ...
, and the Cauchy−Binet formula to prove has been rewritten as : \begin & \sum_\sum_p(f,g)\det(L_fR_g) \\ pt= & \sum_\sum_ p(f,g) \sum_ \det((L_f)_) \det((R_g)_), \end where ''p''(''f'',''g'') denotes the scalar factor \textstyle(\prod_^mA_)(\prod_^mB_). It remains to prove the Cauchy−Binet formula for ''A'' = ''L''''f'' and ''B'' = ''R''''g'', for all ''f'',''g'': 'm''nbsp;→  'n'' For this step 2, if ''f'' fails to be injective then ''L''''f'' and ''L''''f''''R''''g'' both have two identical rows, and if ''g'' fails to be injective then ''R''''g'' and ''L''''f''''R''''g'' both have two identical columns; in either case both sides of the identity are zero. Supposing now that both ''f'' and ''g'' are injective maps 'm''nbsp;→  'n'' the factor \det((L_f)_) on the right is zero unless ''S'' = ''f''( 'm'', while the factor \det((R_g)_) is zero unless ''S'' = ''g''( 'm''. So if the images of ''f'' and ''g'' are different, the right hand side has only null terms, and the left hand side is zero as well since ''L''''f''''R''''g'' has a null row (for ''i'' with f(i)\notin g( ). In the remaining case where the images of ''f'' and ''g'' are the same, say ''f''( 'm'' = ''S'' = ''g''( 'm'', we need to prove that :\det(L_fR_g)=\det((L_f)_)\det((R_g)_).\, Let ''h'' be the unique increasing bijection 'm''nbsp;→ ''S'', and ''π'',''σ'' the permutations of 'm''such that f=h\circ\pi^ and g=h\circ\sigma; then (L_f)_ is the
permutation matrix In mathematics, particularly in matrix theory, a permutation matrix is a square binary matrix that has exactly one entry of 1 in each row and each column with all other entries 0. An permutation matrix can represent a permutation of elements. ...
for , (R_g)_ is the permutation matrix for ''σ'', and ''L''''f''''R''''g'' is the permutation matrix for \pi\circ\sigma, and since the determinant of a permutation matrix equals the
signature A signature (; from , "to sign") is a depiction of someone's name, nickname, or even a simple "X" or other mark that a person writes on documents as a proof of identity and intent. Signatures are often, but not always, Handwriting, handwritt ...
of the permutation, the identity follows from the fact that signatures are multiplicative. Using multi-linearity with respect to both the rows of ''A'' and the columns of ''B'' in the proof is not necessary; one could use just one of them, say the former, and use that a matrix product ''L''''f''''B'' either consists of a permutation of the rows of ''B''''f''( 'm'', 'm''/sub> (if ''f'' is injective), or has at least two equal rows.


Relation to the generalized Kronecker delta

As we have seen, the Cauchy–Binet formula is equivalent to the following: : \det(L_fR_g)=\sum_ \det((L_f)_)\det((R_g)_), where : L_f=\bigl((\delta_)_\bigr) \quad\text \quad R_g=\bigl((\delta_)_\bigr). In terms of generalized Kronecker delta, we can derive the formula equivalent to the Cauchy–Binet formula: : \delta^_ = \sum_ \delta^_ \delta^_.


Geometric interpretations

If ''A'' is a real ''m''×''n'' matrix, then det(''A'' ''A''T) is equal to the square of the ''m''-dimensional volume of the parallelotope spanned in R''n'' by the ''m'' rows of ''A''. Binet's formula states that this is equal to the sum of the squares of the volumes that arise if the parallelepiped is orthogonally projected onto the ''m''-dimensional coordinate planes (of which there are \tbinom nm). In the case ''m'' = 1 the parallelotope is reduced to a single vector and its volume is its length. The above statement then states that the square of the length of a vector is the sum of the squares of its coordinates; this is indeed the case by the definition of that length, which is based on the
Pythagorean theorem In mathematics, the Pythagorean theorem or Pythagoras' theorem is a fundamental relation in Euclidean geometry between the three sides of a right triangle. It states that the area of the square whose side is the hypotenuse (the side opposite t ...
. In
tensor algebra In mathematics, the tensor algebra of a vector space ''V'', denoted ''T''(''V'') or ''T''(''V''), is the algebra over a field, algebra of tensors on ''V'' (of any rank) with multiplication being the tensor product. It is the free algebra on ''V'', ...
, given an
inner product space In mathematics, an inner product space (or, rarely, a Hausdorff pre-Hilbert space) is a real vector space or a complex vector space with an operation called an inner product. The inner product of two vectors in the space is a scalar, ofte ...
V of dimension ''n'', the Cauchy–Binet formula defines an induced inner product on the
exterior algebra In mathematics, the exterior algebra or Grassmann algebra of a vector space V is an associative algebra that contains V, which has a product, called exterior product or wedge product and denoted with \wedge, such that v\wedge v=0 for every vector ...
\wedge^m V, namely:
\langle v_1\wedge\cdots \wedge v_m, w_1\wedge\cdots \wedge w_m\rangle =\det\left( \langle v_i,w_j\rangle \right)_^m .


Generalization

The Cauchy–Binet formula can be extended in a straightforward way to a general formula for the minors of the product of two matrices. Context for the formula is given in the article on minors, but the idea is that both the formula for ordinary
matrix multiplication In mathematics, specifically in linear algebra, matrix multiplication is a binary operation that produces a matrix (mathematics), matrix from two matrices. For matrix multiplication, the number of columns in the first matrix must be equal to the n ...
and the Cauchy–Binet formula for the determinant of the product of two matrices are special cases of the following general statement about the minors of a product of two matrices. Suppose that A is an ''m'' × ''n'' matrix, B is an ''n'' × ''p'' matrix, ''I'' is a
subset In mathematics, a Set (mathematics), set ''A'' is a subset of a set ''B'' if all Element (mathematics), elements of ''A'' are also elements of ''B''; ''B'' is then a superset of ''A''. It is possible for ''A'' and ''B'' to be equal; if they a ...
of with ''k'' elements and ''J'' is a subset of with ''k'' elements. Then : mathbf = \sum_ mathbf mathbf\, where the sum extends over all subsets ''K'' of with ''k'' elements. Note the notation mathbf means the determinant of the matrix formed by taking only the rows of with index in and the columns with index in .


Continuous version

A continuous version of the Cauchy–Binet formula, known as the Andréief identity, appears commonly in random matrix theory. It is stated as follows: let \left\_^ and \left\_^ be two sequences of integrable functions, supported on I. Then :\int_I \cdots \int_I \det \left _(x_k)\right^N \det \left _(x_k)\right^N dx_1 \cdots dx_N = N!\, \det \left int_I f_j(x)g_k(x) dx\right^. Forrester describes how to recover the usual Cauchy–Binet formula as a discretisation of the above identity. It is occasionally called the Andréief- Heine identity, though the credit to Heine sees unhistorical, as pre-2010 sources generally credit only Andréief.


References

* Joel G. Broida & S. Gill Williamson (1989) ''A Comprehensive Introduction to Linear Algebra'', §4.6 Cauchy-Binet theorem, pp 208–14,
Addison-Wesley Addison–Wesley is an American publisher of textbooks and computer literature. It is an imprint of Pearson plc, a global publishing and education company. In addition to publishing books, Addison–Wesley also distributes its technical titles ...
. * Jin Ho Kwak & Sungpyo Hong (2004) ''Linear Algebra'' 2nd edition, Example 2.15 Binet-Cauchy formula, pp 66,7,
Birkhäuser Birkhäuser was a Swiss publisher founded in 1879 by Emil Birkhäuser. It was acquired by Springer Science+Business Media in 1985. Today it is an imprint used by two companies in unrelated fields: * Springer continues to publish science (parti ...
. * I. R. Shafarevich & A. O. Remizov (2012) ''Linear Algebra and Geometry'', §2.9 (p. 68) & §10.5 (p. 377),
Springer Springer or springers may refer to: Publishers * Springer Science+Business Media, aka Springer International Publishing, a worldwide publishing group founded in 1842 in Germany formerly known as Springer-Verlag. ** Springer Nature, a multinationa ...
. * M.L. Mehta (2004) ''Random matrices'', 3rd ed.,
Elsevier Elsevier ( ) is a Dutch academic publishing company specializing in scientific, technical, and medical content. Its products include journals such as ''The Lancet'', ''Cell (journal), Cell'', the ScienceDirect collection of electronic journals, ...
. {{DEFAULTSORT:Cauchy-Binet formula Determinants Augustin-Louis Cauchy