In
mathematics, in particular
linear algebra
Linear algebra is the branch of mathematics concerning linear equations such as:
:a_1x_1+\cdots +a_nx_n=b,
linear maps such as:
:(x_1, \ldots, x_n) \mapsto a_1x_1+\cdots +a_nx_n,
and their representations in vector spaces and through matric ...
, the Sherman–Morrison formula,
named after Jack Sherman and Winifred J. Morrison, computes the inverse of the sum of an
invertible
In mathematics, the concept of an inverse element generalises the concepts of opposite () and reciprocal () of numbers.
Given an operation denoted here , and an identity element denoted , if , one says that is a left inverse of , and that ...
matrix
Matrix most commonly refers to:
* ''The Matrix'' (franchise), an American media franchise
** '' The Matrix'', a 1999 science-fiction action film
** "The Matrix", a fictional setting, a virtual reality environment, within ''The Matrix'' (franchi ...
and the
outer product
In linear algebra, the outer product of two coordinate vectors is a matrix. If the two vectors have dimensions ''n'' and ''m'', then their outer product is an ''n'' × ''m'' matrix. More generally, given two tensors (multidimensional arrays of n ...
,
, of
vectors and
. The Sherman–Morrison formula is a special case of the
Woodbury formula. Though named after Sherman and Morrison, it appeared already in earlier publications.
Statement
Suppose
is an
invertible
In mathematics, the concept of an inverse element generalises the concepts of opposite () and reciprocal () of numbers.
Given an operation denoted here , and an identity element denoted , if , one says that is a left inverse of , and that ...
square matrix
In mathematics, a square matrix is a matrix with the same number of rows and columns. An ''n''-by-''n'' matrix is known as a square matrix of order Any two square matrices of the same order can be added and multiplied.
Square matrices are ofte ...
and
are
column vector
In linear algebra, a column vector with m elements is an m \times 1 matrix consisting of a single column of m entries, for example,
\boldsymbol = \begin x_1 \\ x_2 \\ \vdots \\ x_m \end.
Similarly, a row vector is a 1 \times n matrix for some n, ...
s. Then
is invertible
iff
In logic and related fields such as mathematics and philosophy, "if and only if" (shortened as "iff") is a biconditional logical connective between statements, where either both statements are true or both are false.
The connective is bicondi ...
. In this case,
:
Here,
is the
outer product
In linear algebra, the outer product of two coordinate vectors is a matrix. If the two vectors have dimensions ''n'' and ''m'', then their outer product is an ''n'' × ''m'' matrix. More generally, given two tensors (multidimensional arrays of n ...
of two vectors
and
. The general form shown here is the one published by Bartlett.
[
]
Proof
(
) To prove that the backward direction
is invertible with inverse given as above) is true, we verify the properties of the inverse. A matrix
(in this case the right-hand side of the Sherman–Morrison formula) is the inverse of a matrix
(in this case
) if and only if
.
We first verify that the right hand side (
) satisfies
.
:
To end the proof of this direction, we need to show that
in a similar way as above:
:
(In fact, the last step can be avoided since for square matrices
and
,
is equivalent to
.)
(
) Reciprocally, if
, then via the
matrix determinant lemma,
, so
is not invertible.
Application
If the inverse of
is already known, the formula provides a
numerically
Numerical analysis is the study of algorithms that use numerical approximation (as opposed to symbolic manipulations) for the problems of mathematical analysis (as distinguished from discrete mathematics). It is the study of numerical methods th ...
cheap way to compute the inverse of
corrected by the matrix
(depending on the point of view, the correction may be seen as a
perturbation
Perturbation or perturb may refer to:
* Perturbation theory, mathematical methods that give approximate solutions to problems that cannot be solved exactly
* Perturbation (geology), changes in the nature of alluvial deposits over time
* Perturbatio ...
or as a
rank
Rank is the relative position, value, worth, complexity, power, importance, authority, level, etc. of a person or object within a ranking, such as:
Level or position in a hierarchical organization
* Academic rank
* Diplomatic rank
* Hierarchy
* H ...
-1 update). The computation is relatively cheap because the inverse of
does not have to be computed from scratch (which in general is expensive), but can be computed by correcting (or perturbing)
.
Using unit columns (columns from the
identity matrix
In linear algebra, the identity matrix of size n is the n\times n square matrix with ones on the main diagonal and zeros elsewhere.
Terminology and notation
The identity matrix is often denoted by I_n, or simply by I if the size is immaterial ...
) for
or
, individual columns or rows of
may be manipulated and a correspondingly updated inverse computed relatively cheaply in this way. In the general case, where
is a
-by-
matrix and
and
are arbitrary vectors of dimension
, the whole matrix is updated
and the computation takes
scalar multiplications. If
is a unit column, the computation takes only
scalar multiplications. The same goes if
is a unit column. If both
and
are unit columns, the computation takes only
scalar multiplications.
This formula also has application in theoretical physics. Namely, in quantum field theory, one uses this formula to calculate the propagator of a spin-1 field. The inverse propagator (as it appears in the Lagrangian) has the form
. One uses the Sherman–Morrison formula to calculate the inverse (satisfying certain time-ordering boundary conditions) of the inverse propagator—or simply the (Feynman) propagator—which is needed to perform any perturbative calculation
involving the spin-1 field.
Alternative verification
Following is an alternate verification of the Sherman–Morrison formula using the easily verifiable identity
:
.
Let
:
then
:
.
Substituting
gives
:
Generalization ( Woodbury matrix identity)
Given a square invertible
matrix
, an
matrix
, and a
matrix
, let
be an
matrix such that
. Then, assuming
is invertible, we have
:
See also
* The
matrix determinant lemma performs a rank-1 update to a
determinant
In mathematics, the determinant is a scalar value that is a function of the entries of a square matrix. It characterizes some properties of the matrix and the linear map represented by the matrix. In particular, the determinant is nonzero if ...
.
*
Woodbury matrix identity
*
Quasi-Newton method Quasi-Newton methods are methods used to either find zeroes or local maxima and minima of functions, as an alternative to Newton's method. They can be used if the Jacobian or Hessian is unavailable or is too expensive to compute at every iteration ...
*
Binomial inverse theorem
*
Bunch–Nielsen–Sorensen formula
*
Maxwell stress tensor
The Maxwell stress tensor (named after James Clerk Maxwell) is a symmetric second-order tensor used in classical electromagnetism to represent the interaction between electromagnetic forces and mechanical momentum. In simple situations, such as a ...
contains an application of the Sherman–Morrison formula.
References
External links
*
{{DEFAULTSORT:Sherman-Morrison formula
Linear algebra
Matrix theory