HOME

TheInfoList



OR:

In
linear algebra Linear algebra is the branch of mathematics concerning linear equations such as :a_1x_1+\cdots +a_nx_n=b, linear maps such as :(x_1, \ldots, x_n) \mapsto a_1x_1+\cdots +a_nx_n, and their representations in vector spaces and through matrix (mathemat ...
, the Sherman–Morrison formula, named after Jack Sherman and Winifred J. Morrison, computes the inverse of a "
rank A rank is a position in a hierarchy. It can be formally recognized—for example, cardinal, chief executive officer, general, professor—or unofficial. People Formal ranks * Academic rank * Corporate title * Diplomatic rank * Hierarchy ...
-1 update" to a
matrix Matrix (: matrices or matrixes) or MATRIX may refer to: Science and mathematics * Matrix (mathematics), a rectangular array of numbers, symbols or expressions * Matrix (logic), part of a formula in prenex normal form * Matrix (biology), the m ...
whose inverse has previously been computed. That is, given an
invertible In mathematics, the concept of an inverse element generalises the concepts of opposite () and reciprocal () of numbers. Given an operation denoted here , and an identity element denoted , if , one says that is a left inverse of , and that ...
matrix A and the
outer product In linear algebra, the outer product of two coordinate vectors is the matrix whose entries are all products of an element in the first vector with an element in the second vector. If the two coordinate vectors have dimensions ''n'' and ''m'', the ...
u v^\textsf of
vector Vector most often refers to: * Euclidean vector, a quantity with a magnitude and a direction * Disease vector, an agent that carries and transmits an infectious pathogen into another living organism Vector may also refer to: Mathematics a ...
s u and v, the formula cheaply computes an updated matrix inverse \left(A + uv^\textsf\right)\vphantom)^. The Sherman–Morrison formula is a special case of the Woodbury formula. Though named after Sherman and Morrison, it appeared already in earlier publications.


Statement

Suppose A\in\mathbb^ is an invertible
square matrix In mathematics, a square matrix is a Matrix (mathematics), matrix with the same number of rows and columns. An ''n''-by-''n'' matrix is known as a square matrix of order Any two square matrices of the same order can be added and multiplied. Squ ...
and u,v\in\mathbb^n are
column vector In linear algebra, a column vector with elements is an m \times 1 matrix consisting of a single column of entries, for example, \boldsymbol = \begin x_1 \\ x_2 \\ \vdots \\ x_m \end. Similarly, a row vector is a 1 \times n matrix for some , c ...
s. Then A + uv^\textsf is invertible
if and only if In logic and related fields such as mathematics and philosophy, "if and only if" (often shortened as "iff") is paraphrased by the biconditional, a logical connective between statements. The biconditional is true in two cases, where either bo ...
1 + v^\textsf A^u \neq 0. In this case, :\left(A + uv^\textsf\right)^ = A^ - . Here, uv^\textsf is the outer product of two vectors u and v. The general form shown here is the one published by Bartlett.


Proof

(\Leftarrow) To prove that the backward direction 1 + v^\textsfA^u \neq 0 \Rightarrow A + uv^\textsf is invertible with inverse given as above) is true, we verify the properties of the inverse. A matrix Y (in this case the right-hand side of the Sherman–Morrison formula) is the inverse of a matrix X (in this case A + uv^\textsf) if and only if XY = YX = I. We first verify that the right hand side (Y) satisfies XY = I. :\begin XY &= \left(A + uv^\textsf\right)\left(A^ - \right) \\ pt &= AA^ + uv^\textsfA^ - \\ pt &= I + uv^\textsfA^ - \\ pt &= I + uv^\textsfA^ - \\ pt &= I + uv^\textsfA^ - uv^\textsfA^\\ pt &= I \end To end the proof of this direction, we need to show that YX = I in a similar way as above: :YX = \left(A^ - \right)(A + uv^\textsf) = I. (In fact, the last step can be avoided since for square matrices X and Y, XY = I is equivalent to YX = I.) (\Rightarrow) Reciprocally, if 1 + v^\textsfA^u = 0 , then via the matrix determinant lemma, \det\!\left(A + uv^\textsf\right)=(1 + v^\textsf A^ u) \det(A) = 0, so \left(A + uv^\textsf\right) is not invertible.


Application

If the inverse of A is already known, the formula provides a numerically cheap way to compute the inverse of A corrected by the matrix uv^\textsf (depending on the point of view, the correction may be seen as a perturbation or as a rank-1 update). The computation is relatively cheap because the inverse of A + uv^\textsf does not have to be computed from scratch (which in general is expensive), but can be computed by correcting (or perturbing) A^. Using unit columns (columns from the
identity matrix In linear algebra, the identity matrix of size n is the n\times n square matrix with ones on the main diagonal and zeros elsewhere. It has unique properties, for example when the identity matrix represents a geometric transformation, the obje ...
) for u or v, individual columns or rows of A may be manipulated and a correspondingly updated inverse computed relatively cheaply in this way. In the general case, where A^ is an n-by-n matrix and u and v are arbitrary vectors of dimension n, the whole matrix is updated and the computation takes 3n^2 scalar multiplications. If u is a unit column, the computation takes only 2n^2 scalar multiplications. The same goes if v is a unit column. If both u and v are unit columns, the computation takes only n^2 scalar multiplications. This formula also has application in
theoretical physics Theoretical physics is a branch of physics that employs mathematical models and abstractions of physical objects and systems to rationalize, explain, and predict List of natural phenomena, natural phenomena. This is in contrast to experimental p ...
. Namely, in
quantum field theory In theoretical physics, quantum field theory (QFT) is a theoretical framework that combines Field theory (physics), field theory and the principle of relativity with ideas behind quantum mechanics. QFT is used in particle physics to construct phy ...
, one uses this formula to calculate the propagator of a spin-1 field. The inverse propagator (as it appears in the Lagrangian) has the form A + uv^\textsf. One uses the Sherman–Morrison formula to calculate the inverse (satisfying certain time-ordering boundary conditions) of the inverse propagator—or simply the (Feynman) propagator—which is needed to perform any perturbative calculation involving the spin-1 field. One of the issues with the formula is that little is known about its numerical stability. There are no published results concerning its error bounds. Anecdotal evidence suggests that the Woodbury matrix identity (a generalization of the Sherman–Morrison formula) may diverge even for seemingly benign examples (when both the original and modified matrices are
well-conditioned In numerical analysis, the condition number of a function measures how much the output value of the function can change for a small change in the input argument. This is used to measure how sensitive a function is to changes or errors in the inpu ...
).


Alternative verification

Following is an alternate verification of the Sherman–Morrison formula using the easily verifiable identity : \left(I + wv^\textsf\right)^ = I - \frac. Let :u = Aw, \quad \text \quad A + uv^\textsf = A\left(I + wv^\textsf\right), then : \left(A + uv^\textsf\right)^ = \left(I + wv^\textsf\right)^ A^ = \left(I - \frac\right)A^. Substituting w = A^u gives : \left(A + uv^\textsf\right)^ = \left(I - \frac \right)A^ = A^ - \frac


Generalization ( Woodbury matrix identity)

Given a square invertible n \times n matrix A, an n \times k matrix U, and a k \times n matrix V, let B be an n \times n matrix such that B = A + UV. Then, assuming \left(I_k + VA^ U\right) is invertible, we have :B^ = A^ - A^U \left(I_k + VA^U\right)^ VA^.


See also

* The matrix determinant lemma performs a rank-1 update to a
determinant In mathematics, the determinant is a Scalar (mathematics), scalar-valued function (mathematics), function of the entries of a square matrix. The determinant of a matrix is commonly denoted , , or . Its value characterizes some properties of the ...
. * Woodbury matrix identity * Quasi-Newton method * Binomial inverse theorem * Bunch–Nielsen–Sorensen formula *
Maxwell stress tensor The Maxwell stress tensor (named after James Clerk Maxwell) is a symmetric second-order tensor in three dimensions that is used in classical electromagnetism to represent the interaction between electromagnetic forces and mechanical momentum. In ...
contains an application of the Sherman–Morrison formula.


References


External links

* {{DEFAULTSORT:Sherman-Morrison formula Linear algebra Matrix theory