HOME

TheInfoList



OR:

In
mathematics Mathematics is an area of knowledge that includes the topics of numbers, formulas and related structures, shapes and the spaces in which they are contained, and quantities and their changes. These topics are represented in modern mathematics ...
, in particular
linear algebra Linear algebra is the branch of mathematics concerning linear equations such as: :a_1x_1+\cdots +a_nx_n=b, linear maps such as: :(x_1, \ldots, x_n) \mapsto a_1x_1+\cdots +a_nx_n, and their representations in vector spaces and through matrices. ...
, the Sherman–Morrison formula, named after Jack Sherman and Winifred J. Morrison, computes the inverse of the sum of an
invertible In mathematics, the concept of an inverse element generalises the concepts of opposite () and reciprocal () of numbers. Given an operation denoted here , and an identity element denoted , if , one says that is a left inverse of , and that is ...
matrix Matrix most commonly refers to: * ''The Matrix'' (franchise), an American media franchise ** ''The Matrix'', a 1999 science-fiction action film ** "The Matrix", a fictional setting, a virtual reality environment, within ''The Matrix'' (franchis ...
A and the
outer product In linear algebra, the outer product of two coordinate vector In linear algebra, a coordinate vector is a representation of a vector as an ordered list of numbers (a tuple) that describes the vector in terms of a particular ordered basis. An ea ...
, u v^\textsf, of vectors u and v. The Sherman–Morrison formula is a special case of the Woodbury formula. Though named after Sherman and Morrison, it appeared already in earlier publications.


Statement

Suppose A\in\mathbb^ is an
invertible In mathematics, the concept of an inverse element generalises the concepts of opposite () and reciprocal () of numbers. Given an operation denoted here , and an identity element denoted , if , one says that is a left inverse of , and that is ...
square matrix In mathematics, a square matrix is a matrix with the same number of rows and columns. An ''n''-by-''n'' matrix is known as a square matrix of order Any two square matrices of the same order can be added and multiplied. Square matrices are often ...
and u,v\in\mathbb^n are
column vector In linear algebra, a column vector with m elements is an m \times 1 matrix consisting of a single column of m entries, for example, \boldsymbol = \begin x_1 \\ x_2 \\ \vdots \\ x_m \end. Similarly, a row vector is a 1 \times n matrix for some n, c ...
s. Then A + uv^\textsf is invertible
iff In logic and related fields such as mathematics and philosophy, "if and only if" (shortened as "iff") is a biconditional logical connective between statements, where either both statements are true or both are false. The connective is bicon ...
1 + v^\textsf A^u \neq 0. In this case, :\left(A + uv^\textsf\right)^ = A^ - . Here, uv^\textsf is the
outer product In linear algebra, the outer product of two coordinate vector In linear algebra, a coordinate vector is a representation of a vector as an ordered list of numbers (a tuple) that describes the vector in terms of a particular ordered basis. An ea ...
of two vectors u and v. The general form shown here is the one published by Bartlett.


Proof

(\Leftarrow) To prove that the backward direction 1 + v^\textsfA^u \neq 0 \Rightarrow A + uv^\textsf is invertible with inverse given as above) is true, we verify the properties of the inverse. A matrix Y (in this case the right-hand side of the Sherman–Morrison formula) is the inverse of a matrix X (in this case A + uv^\textsf) if and only if XY = YX = I. We first verify that the right hand side (Y) satisfies XY = I. :\begin XY &= \left(A + uv^\textsf\right)\left(A^ - \right) \\ pt &= AA^ + uv^\textsfA^ - \\ pt &= I + uv^\textsfA^ - \\ pt &= I + uv^\textsfA^ - \\ pt &= I + uv^\textsfA^ - uv^\textsfA^\\ pt \end To end the proof of this direction, we need to show that YX = I in a similar way as above: :YX = \left(A^ - \right)(A + uv^\textsf) = I. (In fact, the last step can be avoided since for square matrices X and Y, XY = I is equivalent to YX = I.) (\Rightarrow) Reciprocally, if 1 + v^\textsfA^u = 0 , then via the
matrix determinant lemma In mathematics, in particular linear algebra, the matrix determinant lemma computes the determinant of the sum of an invertible matrix A and the dyadic product, uvT, of a column vector u and a row vector vT. Statement Suppose A is an invertible s ...
, \det\!\left(A + uv^\textsf\right)=(1 + v^\textsf A^ u) \det(A) = 0, so \left(A + uv^\textsf\right) is not invertible.


Application

If the inverse of A is already known, the formula provides a numerically
cheap Cheap may refer to: *Cheapness * ''Cheap'' (album), debut album from Seasick Steve *Cheap (ward), London, UK *Flatwoods, Kentucky, previously known as Cheap See also *Cheapskate A miser is a person who is reluctant to spend, sometimes to th ...
way to compute the inverse of A corrected by the matrix uv^\textsf (depending on the point of view, the correction may be seen as a
perturbation Perturbation or perturb may refer to: * Perturbation theory, mathematical methods that give approximate solutions to problems that cannot be solved exactly * Perturbation (geology), changes in the nature of alluvial deposits over time * Perturbatio ...
or as a
rank Rank is the relative position, value, worth, complexity, power, importance, authority, level, etc. of a person or object within a ranking, such as: Level or position in a hierarchical organization * Academic rank * Diplomatic rank * Hierarchy * H ...
-1 update). The computation is relatively cheap because the inverse of A + uv^\textsf does not have to be computed from scratch (which in general is expensive), but can be computed by correcting (or perturbing) A^. Using unit columns (columns from the
identity matrix In linear algebra, the identity matrix of size n is the n\times n square matrix with ones on the main diagonal and zeros elsewhere. Terminology and notation The identity matrix is often denoted by I_n, or simply by I if the size is immaterial o ...
) for u or v, individual columns or rows of A may be manipulated and a correspondingly updated inverse computed relatively cheaply in this way. In the general case, where A^ is a n-by-n matrix and u and v are arbitrary vectors of dimension n, the whole matrix is updated and the computation takes 3n^2 scalar multiplications. If u is a unit column, the computation takes only 2n^2 scalar multiplications. The same goes if v is a unit column. If both u and v are unit columns, the computation takes only n^2 scalar multiplications. This formula also has application in theoretical physics. Namely, in quantum field theory, one uses this formula to calculate the propagator of a spin-1 field. The inverse propagator (as it appears in the Lagrangian) has the form \left(A + uv^\textsf\right). One uses the Sherman–Morrison formula to calculate the inverse (satisfying certain time-ordering boundary conditions) of the inverse propagator—or simply the (Feynman) propagator—which is needed to perform any perturbative calculation involving the spin-1 field.


Alternative verification

Following is an alternate verification of the Sherman–Morrison formula using the easily verifiable identity : \left(I + wv^\textsf\right)^ = I - \frac. Let :u = Aw, \quad \text \quad A + uv^\textsf = A\left(I + wv^\textsf\right), then : \left(A + uv^\textsf\right)^ = \left(I + wv^\textsf\right)^ A^ = \left(I - \frac\right)A^. Substituting w = A^u gives : \left(A + uv^\textsf\right)^ = \left(I - \frac \right)A^ = A^ - \frac


Generalization (

Woodbury matrix identity In mathematics (specifically linear algebra), the Woodbury matrix identity, named after Max A. Woodbury, says that the inverse of a rank-''k'' correction of some matrix can be computed by doing a rank-''k'' correction to the inverse of the origina ...
)

Given a square invertible n \times n matrix A, an n \times k matrix U, and a k \times n matrix V, let B be an n \times n matrix such that B = A + UV. Then, assuming \left(I_k + VA^ U\right) is invertible, we have :B^ = A^ - A^U \left(I_k + VA^U\right)^ VA^.


See also

* The
matrix determinant lemma In mathematics, in particular linear algebra, the matrix determinant lemma computes the determinant of the sum of an invertible matrix A and the dyadic product, uvT, of a column vector u and a row vector vT. Statement Suppose A is an invertible s ...
performs a rank-1 update to a
determinant In mathematics, the determinant is a scalar value that is a function of the entries of a square matrix. It characterizes some properties of the matrix and the linear map represented by the matrix. In particular, the determinant is nonzero if and ...
. *
Woodbury matrix identity In mathematics (specifically linear algebra), the Woodbury matrix identity, named after Max A. Woodbury, says that the inverse of a rank-''k'' correction of some matrix can be computed by doing a rank-''k'' correction to the inverse of the origina ...
*
Quasi-Newton method Quasi-Newton methods are methods used to either find zeroes or local maxima and minima of functions, as an alternative to Newton's method. They can be used if the Jacobian or Hessian is unavailable or is too expensive to compute at every iteration. ...
*
Binomial inverse theorem In mathematics (specifically linear algebra), the Woodbury matrix identity, named after Max A. Woodbury, says that the inverse of a rank-''k'' correction of some matrix can be computed by doing a rank-''k'' correction to the inverse of the origina ...
*
Bunch–Nielsen–Sorensen formula In mathematics, in particular linear algebra, the Bunch–Nielsen–Sorensen formula, named after James R. Bunch, Christopher P. Nielsen and Danny C. Sorensen, expresses the eigenvectors of the sum of a symmetric matrix A and the outer product, v v^ ...
*
Maxwell stress tensor The Maxwell stress tensor (named after James Clerk Maxwell) is a symmetric second-order tensor used in classical electromagnetism to represent the interaction between electromagnetic forces and mechanical momentum. In simple situations, such as a ...
contains an application of the Sherman–Morrison formula.


References


External links

* {{DEFAULTSORT:Sherman-Morrison formula Linear algebra Matrix theory