In
statistics
Statistics (from German language, German: ''wikt:Statistik#German, Statistik'', "description of a State (polity), state, a country") is the discipline that concerns the collection, organization, analysis, interpretation, and presentation of ...
and
signal processing
Signal processing is an electrical engineering subfield that focuses on analyzing, modifying and synthesizing ''signals'', such as audio signal processing, sound, image processing, images, and scientific measurements. Signal processing techniq ...
, a minimum mean square error (MMSE) estimator is an estimation method which minimizes the
mean square error (MSE), which is a common measure of estimator quality, of the fitted values of a
dependent variable
Dependent and independent variables are variables in mathematical modeling, statistical modeling and experimental sciences. Dependent variables receive this name because, in an experiment, their values are studied under the supposition or demand ...
. In the
Bayesian
Thomas Bayes (/beɪz/; c. 1701 – 1761) was an English statistician, philosopher, and Presbyterian minister.
Bayesian () refers either to a range of concepts and approaches that relate to statistical methods based on Bayes' theorem, or a followe ...
setting, the term MMSE more specifically refers to estimation with quadratic
loss function. In such case, the MMSE estimator is given by the posterior mean of the parameter to be estimated. Since the posterior mean is cumbersome to calculate, the form of the MMSE estimator is usually constrained to be within a certain class of functions. Linear MMSE estimators are a popular choice since they are easy to use, easy to calculate, and very versatile. It has given rise to many popular estimators such as the
Wiener–Kolmogorov filter and
Kalman filter
For statistics and control theory, Kalman filtering, also known as linear quadratic estimation (LQE), is an algorithm that uses a series of measurements observed over time, including statistical noise and other inaccuracies, and produces estima ...
.
Motivation
The term MMSE more specifically refers to estimation in a
Bayesian
Thomas Bayes (/beɪz/; c. 1701 – 1761) was an English statistician, philosopher, and Presbyterian minister.
Bayesian () refers either to a range of concepts and approaches that relate to statistical methods based on Bayes' theorem, or a followe ...
setting with quadratic cost function. The basic idea behind the Bayesian approach to estimation stems from practical situations where we often have some prior information about the parameter to be estimated. For instance, we may have prior information about the range that the parameter can assume; or we may have an old estimate of the parameter that we want to modify when a new observation is made available; or the statistics of an actual random signal such as speech. This is in contrast to the non-Bayesian approach like
minimum-variance unbiased estimator In statistics a minimum-variance unbiased estimator (MVUE) or uniformly minimum-variance unbiased estimator (UMVUE) is an unbiased estimator that has lower variance than any other unbiased estimator for all possible values of the parameter.
For pra ...
(MVUE) where absolutely nothing is assumed to be known about the parameter in advance and which does not account for such situations. In the Bayesian approach, such prior information is captured by the prior probability density function of the parameters; and based directly on
Bayes theorem
In probability theory and statistics, Bayes' theorem (alternatively Bayes' law or Bayes' rule), named after Thomas Bayes, describes the probability of an event, based on prior knowledge of conditions that might be related to the event. For examp ...
, it allows us to make better posterior estimates as more observations become available. Thus unlike non-Bayesian approach where parameters of interest are assumed to be deterministic, but unknown constants, the Bayesian estimator seeks to estimate a parameter that is itself a random variable. Furthermore, Bayesian estimation can also deal with situations where the sequence of observations are not necessarily independent. Thus Bayesian estimation provides yet another alternative to the MVUE. This is useful when the MVUE does not exist or cannot be found.
Definition
Let
be a
hidden random vector variable, and let
be a
known random vector variable (the measurement or observation), both of them not necessarily of the same dimension. An
estimator
In statistics, an estimator is a rule for calculating an estimate of a given quantity based on observed data: thus the rule (the estimator), the quantity of interest (the estimand) and its result (the estimate) are distinguished. For example, the ...
of
is any function of the measurement
. The estimation error vector is given by
and its
mean squared error
In statistics, the mean squared error (MSE) or mean squared deviation (MSD) of an estimator (of a procedure for estimating an unobserved quantity) measures the average of the squares of the errors—that is, the average squared difference between ...
(MSE) is given by the
trace
Trace may refer to:
Arts and entertainment Music
* ''Trace'' (Son Volt album), 1995
* ''Trace'' (Died Pretty album), 1993
* Trace (band), a Dutch progressive rock band
* ''The Trace'' (album)
Other uses in arts and entertainment
* ''Trace'' ...
of error
covariance matrix
:
where the
expectation is taken over
conditioned on
. When
is a scalar variable, the MSE expression simplifies to
. Note that MSE can equivalently be defined in other ways, since
:
The MMSE estimator is then defined as the estimator achieving minimal MSE:
:
Properties
* When the means and variances are finite, the MMSE estimator is uniquely defined and is given by:
::
:In other words, the MMSE estimator is the conditional expectation of
given the known observed value of the measurements. Also, since
is the posterior mean, the error covariance matrix
is equal to the posterior covariance
matrix,
::
.
* The MMSE estimator is unbiased (under the regularity assumptions mentioned above):
::
* The MMSE estimator is
asymptotically unbiased
In analytic geometry, an asymptote () of a curve is a line such that the distance between the curve and the line approaches zero as one or both of the ''x'' or ''y'' coordinates tends to infinity. In projective geometry and related contexts, ...
and it converges in distribution to the normal distribution:
::
:where
is the
Fisher information
In mathematical statistics, the Fisher information (sometimes simply called information) is a way of measuring the amount of information that an observable random variable ''X'' carries about an unknown parameter ''θ'' of a distribution that model ...
of
. Thus, the MMSE estimator is
asymptotically efficient.
* The
orthogonality principle: When
is a scalar, an estimator constrained to be of certain form
is an optimal estimator, i.e.
if and only if
::
:for all
in closed, linear subspace
of the measurements. For random vectors, since the MSE for estimation of a random vector is the sum of the MSEs of the coordinates, finding the MMSE estimator of a random vector decomposes into finding the MMSE estimators of the coordinates of X separately:
::
:for all ''i'' and ''j''. More succinctly put, the cross-correlation between the minimum estimation error
and the estimator
should be zero,
::
* If
and
are
jointly Gaussian
In probability theory and statistics, the multivariate normal distribution, multivariate Gaussian distribution, or joint normal distribution is a generalization of the one-dimensional (univariate) normal distribution to higher dimensions. One ...
, then the MMSE estimator is linear, i.e., it has the form
for matrix
and constant
. This can be directly shown using the Bayes theorem. As a consequence, to find the MMSE estimator, it is sufficient to find the linear MMSE estimator.
Linear MMSE estimator
In many cases, it is not possible to determine the analytical expression of the MMSE estimator. Two basic numerical approaches to obtain the MMSE estimate depends on either finding the conditional expectation
or finding the minima of MSE. Direct numerical evaluation of the conditional expectation is computationally expensive since it often requires multidimensional integration usually done via
Monte Carlo methods
Monte Carlo methods, or Monte Carlo experiments, are a broad class of computational algorithms that rely on repeated random sampling to obtain numerical results. The underlying concept is to use randomness to solve problems that might be determini ...
. Another computational approach is to directly seek the minima of the MSE using techniques such as the
stochastic gradient descent methods ; but this method still requires the evaluation of expectation. While these numerical methods have been fruitful, a closed form expression for the MMSE estimator is nevertheless possible if we are willing to make some compromises.
One possibility is to abandon the full optimality requirements and seek a technique minimizing the MSE within a particular class of estimators, such as the class of linear estimators. Thus, we postulate that the conditional expectation of
given
is a simple linear function of
,
, where the measurement
is a random vector,
is a matrix and
is a vector. This can be seen as the first order Taylor approximation of
. The linear MMSE estimator is the estimator achieving minimum MSE among all estimators of such form. That is, it solves the following the optimization problem:
:
One advantage of such linear MMSE estimator is that it is not necessary to explicitly calculate the posterior probability density function of
. Such linear estimator only depends on the first two moments of
and
. So although it may be convenient to assume that
and
are jointly Gaussian, it is not necessary to make this assumption, so long as the assumed distribution has well defined first and second moments. The form of the linear estimator does not depend on the type of the assumed underlying distribution.
The expression for optimal
and
is given by:
:
:
where
,
the
is cross-covariance matrix between
and
, the
is auto-covariance matrix of
.
Thus, the expression for linear MMSE estimator, its mean, and its auto-covariance is given by
:
:
:
where the
is cross-covariance matrix between
and
.
Lastly, the error covariance and minimum mean square error achievable by such estimator is
:
:
Let us have the optimal linear MMSE estimator given as
, where we are required to find the expression for
and
. It is required that the MMSE estimator be unbiased. This means,
:
Plugging the expression for
in above, we get
:
where
and
. Thus we can re-write the estimator as
:
and the expression for estimation error becomes
:
From the orthogonality principle, we can have
, where we take
. Here the left-hand-side term is
:
When equated to zero, we obtain the desired expression for
as
:
The
is cross-covariance matrix between X and Y, and
is auto-covariance matrix of Y. Since
, the expression can also be re-written in terms of
as
:
Thus the full expression for the linear MMSE estimator is
:
Since the estimate
is itself a random variable with
, we can also obtain its auto-covariance as
:
Putting the expression for
and
, we get
:
Lastly, the covariance of linear MMSE estimation error will then be given by
:
The first term in the third line is zero due to the orthogonality principle. Since
, we can re-write
in terms of covariance matrices as
:
This we can recognize to be the same as
Thus the minimum mean square error achievable by such a linear estimator is
:
.
Univariate case
For the special case when both
and
are scalars, the above relations simplify to
:
:
where
is the
Pearson's correlation coefficient
In statistics, the Pearson correlation coefficient (PCC, pronounced ) ― also known as Pearson's ''r'', the Pearson product-moment correlation coefficient (PPMCC), the bivariate correlation, or colloquially simply as the correlation coefficient ...
between
and
.
The above two equations allows us to interpret the correlation coefficient either as normalized slope of linear regression
:
or as square root of the ratio of two variances
:
.
When
, we have
and
. In this case, no new information is gleaned from the measurement which can decrease the uncertainty in
. On the other hand, when
, we have
and
. Here
is completely determined by
, as given by the equation of straight line.
Computation
Standard method like
Gauss elimination
In mathematics, Gaussian elimination, also known as row reduction, is an algorithm for solving systems of linear equations. It consists of a sequence of operations performed on the corresponding matrix of coefficients. This method can also be used ...
can be used to solve the matrix equation for
. A more numerically stable method is provided by
QR decomposition
In linear algebra, a QR decomposition, also known as a QR factorization or QU factorization, is a decomposition of a matrix ''A'' into a product ''A'' = ''QR'' of an orthogonal matrix ''Q'' and an upper triangular matrix ''R''. QR decomp ...
method. Since the matrix
is a symmetric positive definite matrix,
can be solved twice as fast with the
Cholesky decomposition
In linear algebra, the Cholesky decomposition or Cholesky factorization (pronounced ) is a decomposition of a Hermitian, positive-definite matrix into the product of a lower triangular matrix and its conjugate transpose, which is useful for effici ...
, while for large sparse systems
conjugate gradient method is more effective.
Levinson recursion Levinson recursion or Levinson–Durbin recursion is a procedure in linear algebra to recursively calculate the solution to an equation involving a Toeplitz matrix. The algorithm runs in time, which is a strong improvement over Gauss–Jordan eli ...
is a fast method when
is also a
Toeplitz matrix In linear algebra, a Toeplitz matrix or diagonal-constant matrix, named after Otto Toeplitz, is a matrix in which each descending diagonal from left to right is constant. For instance, the following matrix is a Toeplitz matrix:
:\qquad\begin
a & b ...
. This can happen when
is a
wide sense stationary
In mathematics and statistics, a stationary process (or a strict/strictly stationary process or strong/strongly stationary process) is a stochastic process whose unconditional joint probability distribution does not change when shifted in time. Con ...
process. In such stationary cases, these estimators are also referred to as
Wiener–Kolmogorov filters.
Linear MMSE estimator for linear observation process
Let us further model the underlying process of observation as a linear process:
, where
is a known matrix and
is random noise vector with the mean
and cross-covariance
. Here the required mean and the covariance matrices will be
:
:
:
Thus the expression for the linear MMSE estimator matrix
further modifies to
:
Putting everything into the expression for
, we get
:
Lastly, the error covariance is
:
The significant difference between the estimation problem treated above and those of
least squares and
Gauss–Markov estimate is that the number of observations ''m'', (i.e. the dimension of
) need not be at least as large as the number of unknowns, ''n'', (i.e. the dimension of
). The estimate for the linear observation process exists so long as the ''m''-by-''m'' matrix
exists; this is the case for any ''m'' if, for instance,
is positive definite. Physically the reason for this property is that since
is now a random variable, it is possible to form a meaningful estimate (namely its mean) even with no measurements. Every new measurement simply provides additional information which may modify our original estimate. Another feature of this estimate is that for ''m'' < ''n'', there need be no measurement error. Thus, we may have
, because as long as
is positive definite, the estimate still exists. Lastly, this technique can handle cases where the noise is correlated.
Alternative form
An alternative form of expression can be obtained by using the matrix identity
:
which can be established by post-multiplying by
and pre-multiplying by
to obtain
:
and
:
Since
can now be written in terms of
as
, we get a simplified expression for
as
:
In this form the above expression can be easily compared with
weighed least square and
Gauss–Markov estimate. In particular, when
, corresponding to infinite variance of the apriori information concerning
, the result
is identical to the weighed linear least square estimate with
as the weight matrix. Moreover, if the components of
are uncorrelated and have equal variance such that
where
is an identity matrix, then
is identical to the ordinary least square estimate.
Sequential linear MMSE estimation
In many real-time applications, observational data is not available in a single batch. Instead the observations are made in a sequence. One possible approach is to use the sequential observations to update an old estimate as additional data becomes available, leading to finer estimates. One crucial difference between batch estimation and sequential estimation is that sequential estimation requires an additional Markov assumption.
In the Bayesian framework, such recursive estimation is easily facilitated using Bayes' rule. Given
observations,
, Bayes' rule gives us the posterior density of
as
:
The
is called the posterior density,
is called the likelihood function, and
is the prior density of ''k''-th time step. Note that the prior density for ''k''-th time step is the posterior density of (''k''-1)-th time step. This structure allows us to formulate a recursive approach to estimation. Here we have assumed the conditional independence of
from previous observations
given
as
:
This is the Markov assumption.
The MMSE estimate
given the
th observation is then the mean of the posterior density
. Here, we have implicitly assumed that the statistical properties of
does not change with time. In other words,
is stationary.
In the context of linear MMSE estimator, the formula for the estimate will have the same form as before. However, the mean and covariance matrices of
and
will need to be replaced by those of the prior density
and likelihood
, respectively.
The mean and covariance matrix of the prior density
is given by the previous MMSE estimate,
, and the error covariance matrix,
:
respectively, as per by the property of MMSE estimators.
Similarly, for the linear observation process, the mean and covariance matrix of the likelihood
is given by
and
:
.
The difference between the predicted value of
, as given by
, and the observed value of
gives the prediction error
, which is also referred to as innovation. It is more convenient to represent the linear MMSE in terms of the prediction error, whose mean and covariance are
and
:
.
Hence, in the estimate update formula, we should replace
and
by
and
, respectively. Also, we should replace
and
by
and
. Lastly, we replace
by
:
Thus, we have the new estimate as
:
and the new error covariance as
:
From the point of view of linear algebra, for sequential estimation, if we have an estimate
based on measurements generating space
, then after receiving another set of measurements, we should subtract out from these measurements that part that could be anticipated from the result of the first measurements. In other words, the updating must be based on that part of the new data which is orthogonal to the old data.
The repeated use of the above two equations as more observations become available lead to recursive estimation techniques. The expressions can be more compactly written as
:
:
:
The matrix
is often referred to as the Kalman gain factor. The alternative formulation of the above algorithm will give
:
:
:
The repetition of these three steps as more data becomes available leads to an iterative estimation algorithm. The generalization of this idea to non-stationary cases gives rise to the
Kalman filter
For statistics and control theory, Kalman filtering, also known as linear quadratic estimation (LQE), is an algorithm that uses a series of measurements observed over time, including statistical noise and other inaccuracies, and produces estima ...
. The three update steps outlined above indeed form the update step of the Kalman filter.
Special case: scalar observations
As an important special case, an easy to use recursive expression can be derived when at each ''k''-th time instant the underlying linear observation process yields a scalar such that
, where
is ''n''-by-1 known column vector whose values can change with time,
is ''n''-by-1 random column vector to be estimated, and
is scalar noise term with variance
. After (''k''+1)-th observation, the direct use of above recursive equations give the expression for the estimate
as:
:
where
is the new scalar observation and the gain factor
is ''n''-by-1 column vector given by
:
The
is ''n''-by-''n'' error covariance matrix given by
:
Here, no matrix inversion is required. Also, the gain factor,
, depends on our confidence in the new data sample, as measured by the noise variance, versus that in the previous data. The initial values of
and
are taken to be the mean and covariance of the aprior probability density function of
.
Alternative approaches: This important special case has also given rise to many other iterative methods (or
adaptive filter
An adaptive filter is a system with a linear filter that has a transfer function controlled by variable parameters and a means to adjust those parameters according to an optimization algorithm. Because of the complexity of the optimization algorit ...
s), such as the
least mean squares filter
Least mean squares (LMS) algorithms are a class of adaptive filter used to mimic a desired filter by finding the filter coefficients that relate to producing the least mean square of the error signal (difference between the desired and the actual ...
and
recursive least squares filter Recursive least squares (RLS) is an adaptive filter algorithm that recursively finds the coefficients that minimize a weighted linear least squares cost function relating to the input signals. This approach is in contrast to other algorithms such ...
, that directly solves the original MSE optimization problem using
stochastic gradient descents. However, since the estimation error
cannot be directly observed, these methods try to minimize the mean squared prediction error
. For instance, in the case of scalar observations, we have the gradient
Thus, the update equation for the least mean square filter is given by
:
where
is the scalar step size and the expectation is approximated by the instantaneous value
. As we can see, these methods bypass the need for covariance matrices.
Examples
Example 1
We shall take a
linear prediction
Linear prediction is a mathematical operation where future values of a discrete-time signal are estimated as a linear function of previous samples.
In digital signal processing, linear prediction is often called linear predictive coding (LPC) and ...
problem as an example. Let a linear combination of observed scalar random variables
and
be used to estimate another future scalar random variable
such that
. If the random variables
are real Gaussian random variables with zero mean and its covariance matrix given by
:
then our task is to find the coefficients
such that it will yield an optimal linear estimate
.
In terms of the terminology developed in the previous sections, for this problem we have the observation vector
, the estimator matrix