In
mathematics
Mathematics is an area of knowledge that includes the topics of numbers, formulas and related structures, shapes and the spaces in which they are contained, and quantities and their changes. These topics are represented in modern mathematics ...
, in particular
linear algebra
Linear algebra is the branch of mathematics concerning linear equations such as:
:a_1x_1+\cdots +a_nx_n=b,
linear maps such as:
:(x_1, \ldots, x_n) \mapsto a_1x_1+\cdots +a_nx_n,
and their representations in vector spaces and through matrices.
...
, the Sherman–Morrison formula,
named after Jack Sherman and Winifred J. Morrison, computes the inverse of the sum of an
invertible
In mathematics, the concept of an inverse element generalises the concepts of opposite () and reciprocal () of numbers.
Given an operation denoted here , and an identity element denoted , if , one says that is a left inverse of , and that is ...
matrix
Matrix most commonly refers to:
* ''The Matrix'' (franchise), an American media franchise
** ''The Matrix'', a 1999 science-fiction action film
** "The Matrix", a fictional setting, a virtual reality environment, within ''The Matrix'' (franchis ...
and the
outer product
In linear algebra, the outer product of two coordinate vector
In linear algebra, a coordinate vector is a representation of a vector as an ordered list of numbers (a tuple) that describes the vector in terms of a particular ordered basis. An ea ...
,
, of
vectors and
. The Sherman–Morrison formula is a special case of the
Woodbury formula. Though named after Sherman and Morrison, it appeared already in earlier publications.
Statement
Suppose
is an
invertible
In mathematics, the concept of an inverse element generalises the concepts of opposite () and reciprocal () of numbers.
Given an operation denoted here , and an identity element denoted , if , one says that is a left inverse of , and that is ...
square matrix
In mathematics, a square matrix is a matrix with the same number of rows and columns. An ''n''-by-''n'' matrix is known as a square matrix of order Any two square matrices of the same order can be added and multiplied.
Square matrices are often ...
and
are
column vector
In linear algebra, a column vector with m elements is an m \times 1 matrix consisting of a single column of m entries, for example,
\boldsymbol = \begin x_1 \\ x_2 \\ \vdots \\ x_m \end.
Similarly, a row vector is a 1 \times n matrix for some n, c ...
s. Then
is invertible
iff
In logic and related fields such as mathematics and philosophy, "if and only if" (shortened as "iff") is a biconditional logical connective between statements, where either both statements are true or both are false.
The connective is bicon ...
. In this case,
:
Here,
is the
outer product
In linear algebra, the outer product of two coordinate vector
In linear algebra, a coordinate vector is a representation of a vector as an ordered list of numbers (a tuple) that describes the vector in terms of a particular ordered basis. An ea ...
of two vectors
and
. The general form shown here is the one published by Bartlett.
[
]
Proof
(
) To prove that the backward direction
is invertible with inverse given as above) is true, we verify the properties of the inverse. A matrix
(in this case the right-hand side of the Sherman–Morrison formula) is the inverse of a matrix
(in this case
) if and only if
.
We first verify that the right hand side (
) satisfies
.
:
To end the proof of this direction, we need to show that
in a similar way as above:
:
(In fact, the last step can be avoided since for square matrices
and
,
is equivalent to
.)
(
) Reciprocally, if
, then via the
matrix determinant lemma In mathematics, in particular linear algebra, the matrix determinant lemma computes the determinant of the sum of an invertible matrix A and the dyadic product, uvT, of a column vector u and a row vector vT.
Statement
Suppose A is an invertible s ...
,
, so
is not invertible.
Application
If the inverse of
is already known, the formula provides a
numerically cheap
Cheap may refer to:
*Cheapness
* ''Cheap'' (album), debut album from Seasick Steve
*Cheap (ward), London, UK
*Flatwoods, Kentucky, previously known as Cheap
See also
*Cheapskate
A miser is a person who is reluctant to spend, sometimes to th ...
way to compute the inverse of
corrected by the matrix
(depending on the point of view, the correction may be seen as a
perturbation
Perturbation or perturb may refer to:
* Perturbation theory, mathematical methods that give approximate solutions to problems that cannot be solved exactly
* Perturbation (geology), changes in the nature of alluvial deposits over time
* Perturbatio ...
or as a
rank
Rank is the relative position, value, worth, complexity, power, importance, authority, level, etc. of a person or object within a ranking, such as:
Level or position in a hierarchical organization
* Academic rank
* Diplomatic rank
* Hierarchy
* H ...
-1 update). The computation is relatively cheap because the inverse of
does not have to be computed from scratch (which in general is expensive), but can be computed by correcting (or perturbing)
.
Using unit columns (columns from the
identity matrix
In linear algebra, the identity matrix of size n is the n\times n square matrix with ones on the main diagonal and zeros elsewhere.
Terminology and notation
The identity matrix is often denoted by I_n, or simply by I if the size is immaterial o ...
) for
or
, individual columns or rows of
may be manipulated and a correspondingly updated inverse computed relatively cheaply in this way. In the general case, where
is a
-by-
matrix and
and
are arbitrary vectors of dimension
, the whole matrix is updated
and the computation takes
scalar multiplications. If
is a unit column, the computation takes only
scalar multiplications. The same goes if
is a unit column. If both
and
are unit columns, the computation takes only
scalar multiplications.
This formula also has application in theoretical physics. Namely, in quantum field theory, one uses this formula to calculate the propagator of a spin-1 field. The inverse propagator (as it appears in the Lagrangian) has the form
. One uses the Sherman–Morrison formula to calculate the inverse (satisfying certain time-ordering boundary conditions) of the inverse propagator—or simply the (Feynman) propagator—which is needed to perform any perturbative calculation
involving the spin-1 field.
Alternative verification
Following is an alternate verification of the Sherman–Morrison formula using the easily verifiable identity
:
.
Let
:
then
:
.
Substituting
gives
:
Generalization (
Woodbury matrix identity
In mathematics (specifically linear algebra), the Woodbury matrix identity, named after Max A. Woodbury, says that the inverse of a rank-''k'' correction of some matrix can be computed by doing a rank-''k'' correction to the inverse of the origina ...
)
Given a square invertible
matrix
, an
matrix
, and a
matrix
, let
be an
matrix such that
. Then, assuming
is invertible, we have
:
See also
* The
matrix determinant lemma In mathematics, in particular linear algebra, the matrix determinant lemma computes the determinant of the sum of an invertible matrix A and the dyadic product, uvT, of a column vector u and a row vector vT.
Statement
Suppose A is an invertible s ...
performs a rank-1 update to a
determinant
In mathematics, the determinant is a scalar value that is a function of the entries of a square matrix. It characterizes some properties of the matrix and the linear map represented by the matrix. In particular, the determinant is nonzero if and ...
.
*
Woodbury matrix identity
In mathematics (specifically linear algebra), the Woodbury matrix identity, named after Max A. Woodbury, says that the inverse of a rank-''k'' correction of some matrix can be computed by doing a rank-''k'' correction to the inverse of the origina ...
*
Quasi-Newton method Quasi-Newton methods are methods used to either find zeroes or local maxima and minima of functions, as an alternative to Newton's method. They can be used if the Jacobian or Hessian is unavailable or is too expensive to compute at every iteration. ...
*
Binomial inverse theorem
In mathematics (specifically linear algebra), the Woodbury matrix identity, named after Max A. Woodbury, says that the inverse of a rank-''k'' correction of some matrix can be computed by doing a rank-''k'' correction to the inverse of the origina ...
*
Bunch–Nielsen–Sorensen formula In mathematics, in particular linear algebra, the Bunch–Nielsen–Sorensen formula, named after James R. Bunch, Christopher P. Nielsen and Danny C. Sorensen, expresses the eigenvectors of the sum of a symmetric matrix A and the outer product, v v^ ...
*
Maxwell stress tensor
The Maxwell stress tensor (named after James Clerk Maxwell) is a symmetric second-order tensor used in classical electromagnetism to represent the interaction between electromagnetic forces and mechanical momentum. In simple situations, such as a ...
contains an application of the Sherman–Morrison formula.
References
External links
*
{{DEFAULTSORT:Sherman-Morrison formula
Linear algebra
Matrix theory