Minimum mean square error
   HOME

TheInfoList



OR:

In
statistics Statistics (from German: '' Statistik'', "description of a state, a country") is the discipline that concerns the collection, organization, analysis, interpretation, and presentation of data. In applying statistics to a scientific, indust ...
and
signal processing Signal processing is an electrical engineering subfield that focuses on analyzing, modifying and synthesizing '' signals'', such as sound, images, and scientific measurements. Signal processing techniques are used to optimize transmissions, ...
, a minimum mean square error (MMSE) estimator is an estimation method which minimizes the
mean square error In statistics, the mean squared error (MSE) or mean squared deviation (MSD) of an estimator (of a procedure for estimating an unobserved quantity) measures the average of the squares of the errors—that is, the average squared difference between ...
(MSE), which is a common measure of estimator quality, of the fitted values of a
dependent variable Dependent and independent variables are variables in mathematical modeling, statistical modeling and experimental sciences. Dependent variables receive this name because, in an experiment, their values are studied under the supposition or dema ...
. In the Bayesian setting, the term MMSE more specifically refers to estimation with quadratic
loss function In mathematical optimization and decision theory, a loss function or cost function (sometimes also called an error function) is a function that maps an event or values of one or more variables onto a real number intuitively representing some "cos ...
. In such case, the MMSE estimator is given by the posterior mean of the parameter to be estimated. Since the posterior mean is cumbersome to calculate, the form of the MMSE estimator is usually constrained to be within a certain class of functions. Linear MMSE estimators are a popular choice since they are easy to use, easy to calculate, and very versatile. It has given rise to many popular estimators such as the Wiener–Kolmogorov filter and Kalman filter.


Motivation

The term MMSE more specifically refers to estimation in a Bayesian setting with quadratic cost function. The basic idea behind the Bayesian approach to estimation stems from practical situations where we often have some prior information about the parameter to be estimated. For instance, we may have prior information about the range that the parameter can assume; or we may have an old estimate of the parameter that we want to modify when a new observation is made available; or the statistics of an actual random signal such as speech. This is in contrast to the non-Bayesian approach like minimum-variance unbiased estimator (MVUE) where absolutely nothing is assumed to be known about the parameter in advance and which does not account for such situations. In the Bayesian approach, such prior information is captured by the prior probability density function of the parameters; and based directly on
Bayes theorem In probability theory and statistics, Bayes' theorem (alternatively Bayes' law or Bayes' rule), named after Thomas Bayes, describes the probability of an event, based on prior knowledge of conditions that might be related to the event. For exa ...
, it allows us to make better posterior estimates as more observations become available. Thus unlike non-Bayesian approach where parameters of interest are assumed to be deterministic, but unknown constants, the Bayesian estimator seeks to estimate a parameter that is itself a random variable. Furthermore, Bayesian estimation can also deal with situations where the sequence of observations are not necessarily independent. Thus Bayesian estimation provides yet another alternative to the MVUE. This is useful when the MVUE does not exist or cannot be found.


Definition

Let x be a n \times 1 hidden random vector variable, and let y be a m \times 1 known random vector variable (the measurement or observation), both of them not necessarily of the same dimension. An estimator \hat(y) of x is any function of the measurement y. The estimation error vector is given by e = \hat - x and its mean squared error (MSE) is given by the trace of error
covariance matrix In probability theory and statistics, a covariance matrix (also known as auto-covariance matrix, dispersion matrix, variance matrix, or variance–covariance matrix) is a square matrix giving the covariance between each pair of elements of ...
: \operatorname = \operatorname \left\ = \operatorname\, where the expectation \operatorname is taken over x conditioned on y. When x is a scalar variable, the MSE expression simplifies to \operatorname \left\. Note that MSE can equivalently be defined in other ways, since :\operatorname \left\ = \operatorname \left\ = \operatorname\ = \sum_^n \operatorname\. The MMSE estimator is then defined as the estimator achieving minimal MSE: :\hat_(y) = \operatorname_ \operatorname.


Properties

* When the means and variances are finite, the MMSE estimator is uniquely defined and is given by: ::\hat_(y) = \operatorname \. :In other words, the MMSE estimator is the conditional expectation of x given the known observed value of the measurements. Also, since \hat_ is the posterior mean, the error covariance matrix C_eis equal to the posterior covariance C_ matrix, ::C_e = C_. * The MMSE estimator is unbiased (under the regularity assumptions mentioned above): ::\operatorname\ = \operatorname\ = \operatorname\. * The MMSE estimator is
asymptotically unbiased In analytic geometry, an asymptote () of a curve is a line such that the distance between the curve and the line approaches zero as one or both of the ''x'' or ''y'' coordinates tends to infinity. In projective geometry and related contexts, ...
and it converges in distribution to the normal distribution: :: \sqrt(\hat_ - x) \xrightarrow \mathcal\left(0 , I^(x)\right), :where I(x) is the Fisher information of x. Thus, the MMSE estimator is asymptotically efficient. * The
orthogonality principle In statistics and signal processing, the orthogonality principle is a necessary and sufficient condition for the optimality of a Bayesian estimator. Loosely stated, the orthogonality principle says that the error vector of the optimal estimator (i ...
: When x is a scalar, an estimator constrained to be of certain form \hat=g(y) is an optimal estimator, i.e. \hat_=g^*(y), if and only if ::\operatorname \ = 0 :for all g(y) in closed, linear subspace \mathcal = \ of the measurements. For random vectors, since the MSE for estimation of a random vector is the sum of the MSEs of the coordinates, finding the MMSE estimator of a random vector decomposes into finding the MMSE estimators of the coordinates of X separately: ::\operatorname \ = 0, :for all ''i'' and ''j''. More succinctly put, the cross-correlation between the minimum estimation error \hat_-x and the estimator \hat should be zero, ::\operatorname \ = 0. * If x and y are
jointly Gaussian In probability theory and statistics, the multivariate normal distribution, multivariate Gaussian distribution, or joint normal distribution is a generalization of the one-dimensional (univariate) normal distribution to higher dimensions. One d ...
, then the MMSE estimator is linear, i.e., it has the form Wy+b for matrix W and constant b. This can be directly shown using the Bayes theorem. As a consequence, to find the MMSE estimator, it is sufficient to find the linear MMSE estimator.


Linear MMSE estimator

In many cases, it is not possible to determine the analytical expression of the MMSE estimator. Two basic numerical approaches to obtain the MMSE estimate depends on either finding the conditional expectation \operatorname\ or finding the minima of MSE. Direct numerical evaluation of the conditional expectation is computationally expensive since it often requires multidimensional integration usually done via Monte Carlo methods. Another computational approach is to directly seek the minima of the MSE using techniques such as the stochastic gradient descent methods ; but this method still requires the evaluation of expectation. While these numerical methods have been fruitful, a closed form expression for the MMSE estimator is nevertheless possible if we are willing to make some compromises. One possibility is to abandon the full optimality requirements and seek a technique minimizing the MSE within a particular class of estimators, such as the class of linear estimators. Thus, we postulate that the conditional expectation of x given y is a simple linear function of y, \operatorname\ = W y + b, where the measurement y is a random vector, W is a matrix and b is a vector. This can be seen as the first order Taylor approximation of \operatorname\. The linear MMSE estimator is the estimator achieving minimum MSE among all estimators of such form. That is, it solves the following the optimization problem: :\min_ \operatorname \qquad \text \qquad \hat = W y + b. One advantage of such linear MMSE estimator is that it is not necessary to explicitly calculate the posterior probability density function of x. Such linear estimator only depends on the first two moments of x and y. So although it may be convenient to assume that x and y are jointly Gaussian, it is not necessary to make this assumption, so long as the assumed distribution has well defined first and second moments. The form of the linear estimator does not depend on the type of the assumed underlying distribution. The expression for optimal b and W is given by: :b = \bar - W \bar, : W = C_C^_. where \bar = \operatorname\, \bar = \operatorname\, the C_ is cross-covariance matrix between x and y, the C_ is auto-covariance matrix of y. Thus, the expression for linear MMSE estimator, its mean, and its auto-covariance is given by :\hat = C_C^_(y-\bar) + \bar, :\operatorname\ = \bar, :C_ = C_ C^_Y C_, where the C_ is cross-covariance matrix between y and x. Lastly, the error covariance and minimum mean square error achievable by such estimator is :C_e = C_X - C_ = C_X - C_ C^_Y C_, :\operatorname = \operatorname \. Let us have the optimal linear MMSE estimator given as \hat = Wy+b, where we are required to find the expression for W and b. It is required that the MMSE estimator be unbiased. This means, :\operatorname\ = \operatorname\. Plugging the expression for \hat in above, we get :b = \bar - W \bar, where \bar = \operatorname\ and \bar = \operatorname\. Thus we can re-write the estimator as :\hat = W(y-\bar) + \bar and the expression for estimation error becomes :\hat - x = W(y-\bar) - (x-\bar). From the orthogonality principle, we can have \operatorname \ = 0, where we take g(y) = y - \bar. Here the left-hand-side term is : \begin \operatorname \ &= \operatorname \ \\ &= W \operatorname \ - \operatorname \ \\ &= WC_Y - C_. \end When equated to zero, we obtain the desired expression for W as : W = C_C^_Y . The C_ is cross-covariance matrix between X and Y, and C_ is auto-covariance matrix of Y. Since C_=C^T_, the expression can also be re-written in terms of C_ as :W^T = C^_Y C_ . Thus the full expression for the linear MMSE estimator is :\hat = C_ C^_Y (y-\bar) + \bar. Since the estimate \hat is itself a random variable with \operatorname\ = \bar, we can also obtain its auto-covariance as : \begin C_ &= \operatorname\ \\ &= W \operatorname\ W^T \\ &= W C_Y W^T .\\ \end Putting the expression for W and W^T, we get :C_ = C_ C^_Y C_. Lastly, the covariance of linear MMSE estimation error will then be given by : \begin C_e &= \operatorname\ \\ &= \operatorname\ \\ &= \underbrace_0 W^T - \operatorname\ \\ &= - \operatorname\ \\ &= \operatorname\ - W \operatorname\ \\ &= C_X - WC_ . \end The first term in the third line is zero due to the orthogonality principle. Since W = C_C^_Y, we can re-write C_e in terms of covariance matrices as :C_e = C_X - C_ C^_Y C_ . This we can recognize to be the same as C_e = C_X - C_. Thus the minimum mean square error achievable by such a linear estimator is :\operatorname = \operatorname\ .


Univariate case

For the special case when both x and y are scalars, the above relations simplify to : \hat = \frac(y-\bar) + \bar = \rho \frac(y-\bar) + \bar, :\sigma^2_e = \sigma_X^2 - \frac = (1 - \rho^2)\sigma_X^2, where \rho = \frac is the Pearson's correlation coefficient between x and y. The above two equations allows us to interpret the correlation coefficient either as normalized slope of linear regression : \left(\frac\right) = \rho \left(\frac\right) or as square root of the ratio of two variances :\rho^2 = \frac = \frac. When \rho = 0, we have \hat = \bar and \sigma^2_e = \sigma_X^2. In this case, no new information is gleaned from the measurement which can decrease the uncertainty in x. On the other hand, when \rho = \pm 1, we have \hat = \frac(y-\bar) + \bar and \sigma^2_e = 0. Here x is completely determined by y, as given by the equation of straight line.


Computation

Standard method like Gauss elimination can be used to solve the matrix equation for W. A more numerically stable method is provided by QR decomposition method. Since the matrix C_Y is a symmetric positive definite matrix, W can be solved twice as fast with the Cholesky decomposition, while for large sparse systems
conjugate gradient method In mathematics, the conjugate gradient method is an algorithm In mathematics and computer science, an algorithm () is a finite sequence of rigorous instructions, typically used to solve a class of specific problems or to perform a c ...
is more effective. Levinson recursion is a fast method when C_Y is also a
Toeplitz matrix In linear algebra, a Toeplitz matrix or diagonal-constant matrix, named after Otto Toeplitz, is a matrix in which each descending diagonal from left to right is constant. For instance, the following matrix is a Toeplitz matrix: :\qquad\begin a & b ...
. This can happen when y is a wide sense stationary process. In such stationary cases, these estimators are also referred to as Wiener–Kolmogorov filters.


Linear MMSE estimator for linear observation process

Let us further model the underlying process of observation as a linear process: y=Ax+z, where A is a known matrix and z is random noise vector with the mean \operatorname\=0 and cross-covariance C_ = 0. Here the required mean and the covariance matrices will be :\operatorname\ = A\bar, :C_Y = AC_XA^T + C_Z, :C_ = C_X A^T . Thus the expression for the linear MMSE estimator matrix W further modifies to :W = C_X A^T(AC_XA^T + C_Z)^ . Putting everything into the expression for \hat, we get :\hat = C_X A^T(AC_XA^T + C_Z)^(y-A\bar) + \bar. Lastly, the error covariance is :C_e = C_X - C_ = C_X - C_X A^T(AC_XA^T + C_Z)^AC_X . The significant difference between the estimation problem treated above and those of
least squares The method of least squares is a standard approach in regression analysis to approximate the solution of overdetermined systems (sets of equations in which there are more equations than unknowns) by minimizing the sum of the squares of the re ...
and Gauss–Markov estimate is that the number of observations ''m'', (i.e. the dimension of y) need not be at least as large as the number of unknowns, ''n'', (i.e. the dimension of x). The estimate for the linear observation process exists so long as the ''m''-by-''m'' matrix (AC_XA^T + C_Z)^ exists; this is the case for any ''m'' if, for instance, C_Z is positive definite. Physically the reason for this property is that since x is now a random variable, it is possible to form a meaningful estimate (namely its mean) even with no measurements. Every new measurement simply provides additional information which may modify our original estimate. Another feature of this estimate is that for ''m'' < ''n'', there need be no measurement error. Thus, we may have C_Z = 0, because as long as AC_XA^T is positive definite, the estimate still exists. Lastly, this technique can handle cases where the noise is correlated.


Alternative form

An alternative form of expression can be obtained by using the matrix identity :C_X A^T(AC_XA^T + C_Z)^ = (A^TC_Z^A + C_X^)^ A^T C_Z^, which can be established by post-multiplying by (AC_XA^T + C_Z) and pre-multiplying by (A^TC_Z^A + C_X^), to obtain :W = (A^TC_Z^A + C_X^)^ A^TC_Z^, and :C_e = (A^TC_Z^A + C_X^)^. Since W can now be written in terms of C_e as W = C_e A^T C_Z^, we get a simplified expression for \hat as :\hat = C_e A^T C_Z^(y-A\bar) + \bar. In this form the above expression can be easily compared with weighed least square and Gauss–Markov estimate. In particular, when C_X^=0, corresponding to infinite variance of the apriori information concerning x, the result W = (A^TC_Z^A)^ A^TC_Z^ is identical to the weighed linear least square estimate with C_Z^ as the weight matrix. Moreover, if the components of z are uncorrelated and have equal variance such that C_Z = \sigma^2 I, where I is an identity matrix, then W = (A^TA)^A^T is identical to the ordinary least square estimate.


Sequential linear MMSE estimation

In many real-time applications, observational data is not available in a single batch. Instead the observations are made in a sequence. One possible approach is to use the sequential observations to update an old estimate as additional data becomes available, leading to finer estimates. One crucial difference between batch estimation and sequential estimation is that sequential estimation requires an additional Markov assumption. In the Bayesian framework, such recursive estimation is easily facilitated using Bayes' rule. Given k observations, y_1, \ldots, y_k, Bayes' rule gives us the posterior density of x as : \begin p(x, y_1, \ldots, y_k) &\propto p(y_k, x,y_1,\ldots,y_) p(x, y_1, \ldots, y_) \\ &= p(y_k, x) p(x, y_1, \ldots, y_). \end The p(x, y_1, \ldots, y_k) is called the posterior density, p(y_k, x) is called the likelihood function, and p(x, y_1, \ldots, y_) is the prior density of ''k''-th time step. Note that the prior density for ''k''-th time step is the posterior density of (''k''-1)-th time step. This structure allows us to formulate a recursive approach to estimation. Here we have assumed the conditional independence of y_k from previous observations y_1, \ldots, y_ given x as :p(y_k, x,y_1,\ldots,y_) = p(y_k, x). This is the Markov assumption. The MMSE estimate \hat_k given the kth observation is then the mean of the posterior density p(x, y_1,\ldots, y_k). Here, we have implicitly assumed that the statistical properties of x does not change with time. In other words, x is stationary. In the context of linear MMSE estimator, the formula for the estimate will have the same form as before. However, the mean and covariance matrices of X and Y will need to be replaced by those of the prior density p(x, y_1,\ldots, y_) and likelihood p(y_k, x), respectively. The mean and covariance matrix of the prior density p(x, y_1, \ldots, y_) is given by the previous MMSE estimate, \bar_=\hat_, and the error covariance matrix, :C_ = \mathrm x - \hat_)(x - \hat_)^T= C_, respectively, as per by the property of MMSE estimators. Similarly, for the linear observation process, the mean and covariance matrix of the likelihood p(y_k, x) is given by \bar_k = A\hat_ and : \begin C_ &= \mathrm y_k-\bar_k)(y_k-\bar_k)^T\\ &= \mathrm A(x-\bar_) + z)(A(x-\bar_) + z)^T\\ &= A C_ A^T + C_Z. \end . The difference between the predicted value of y_k, as given by \bar_k = A\hat_, and the observed value of y_k gives the prediction error \tilde_k = y_k - \bar_k, which is also referred to as innovation. It is more convenient to represent the linear MMSE in terms of the prediction error, whose mean and covariance are \mathrm tilde_k= 0 and :C_ = C_. Hence, in the estimate update formula, we should replace \bar and C_X by \hat_ and C_, respectively. Also, we should replace \bar and C_Y by \bar_ and C_. Lastly, we replace C_ by : \begin C_ &= \mathrm x-\hat_)(y_k-\bar_k)^T\\ &= \mathrm x-\hat_)(A(x-\hat_)+z)^T\\ &= C_A^T. \end Thus, we have the new estimate as : \begin \hat_k &= \hat_ + C_ C_^ (y_k - \bar_k) \\ &= \hat_ + C_A^T (AC_A^T + C_Z)^(y_k-A\hat_) \end and the new error covariance as :C_ = C_ - C_A^T(AC_A^T + C_Z)^AC_. From the point of view of linear algebra, for sequential estimation, if we have an estimate \hat_1 based on measurements generating space Y_1, then after receiving another set of measurements, we should subtract out from these measurements that part that could be anticipated from the result of the first measurements. In other words, the updating must be based on that part of the new data which is orthogonal to the old data. The repeated use of the above two equations as more observations become available lead to recursive estimation techniques. The expressions can be more compactly written as :W_ = C_ A^T(AC_A^T + C_Z)^, :\hat_ = \hat_ + W_ (y_-A\hat_), :C_ = (I - W_ A)C_. The matrix W_k is often referred to as the Kalman gain factor. The alternative formulation of the above algorithm will give :C_^ = C_^ + A^T C_Z^ A, :W_ = C_ A^T C_Z^, :\hat_ = \hat_ + W_ (y_-A\hat_), The repetition of these three steps as more data becomes available leads to an iterative estimation algorithm. The generalization of this idea to non-stationary cases gives rise to the Kalman filter. The three update steps outlined above indeed form the update step of the Kalman filter.


Special case: scalar observations

As an important special case, an easy to use recursive expression can be derived when at each ''k''-th time instant the underlying linear observation process yields a scalar such that y_k = a_k^T x_k + z_k, where a_k is ''n''-by-1 known column vector whose values can change with time, x_k is ''n''-by-1 random column vector to be estimated, and z_k is scalar noise term with variance \sigma_k^2. After (''k''+1)-th observation, the direct use of above recursive equations give the expression for the estimate \hat_ as: :\hat_ = \hat_k + w_(y_ - a^T_ \hat_k) where y_ is the new scalar observation and the gain factor w_ is ''n''-by-1 column vector given by :w_ = \frac. The C_ is ''n''-by-''n'' error covariance matrix given by :C_ = (I - w_a^T_)C_ . Here, no matrix inversion is required. Also, the gain factor, w_, depends on our confidence in the new data sample, as measured by the noise variance, versus that in the previous data. The initial values of \hat and C_e are taken to be the mean and covariance of the aprior probability density function of x. Alternative approaches: This important special case has also given rise to many other iterative methods (or
adaptive filter An adaptive filter is a system with a linear filter that has a transfer function controlled by variable parameters and a means to adjust those parameters according to an optimization algorithm. Because of the complexity of the optimization algor ...
s), such as the least mean squares filter and recursive least squares filter, that directly solves the original MSE optimization problem using
stochastic gradient descent Stochastic gradient descent (often abbreviated SGD) is an iterative method for optimizing an objective function with suitable smoothness properties (e.g. differentiable or subdifferentiable). It can be regarded as a stochastic approximation of ...
s. However, since the estimation error e cannot be directly observed, these methods try to minimize the mean squared prediction error \mathrm\. For instance, in the case of scalar observations, we have the gradient \nabla_ \mathrm\ = -2 \mathrm\. Thus, the update equation for the least mean square filter is given by :\hat_ = \hat_k + \eta_k \mathrm\, where \eta_k is the scalar step size and the expectation is approximated by the instantaneous value \mathrm\ \approx a_k \tilde_k. As we can see, these methods bypass the need for covariance matrices.


Examples


Example 1

We shall take a linear prediction problem as an example. Let a linear combination of observed scalar random variables z_, z_ and z_ be used to estimate another future scalar random variable z_ such that \hat z_=\sum_^w_z_. If the random variables z= _,z_,z_,z_ are real Gaussian random variables with zero mean and its covariance matrix given by : \operatorname(Z)=\operatorname z^\left begin 1 & 2 & 3 & 4\\ 2 & 5 & 8 & 9\\ 3 & 8 & 6 & 10\\ 4 & 9 & 10 & 15\end\right then our task is to find the coefficients w_ such that it will yield an optimal linear estimate \hat z_. In terms of the terminology developed in the previous sections, for this problem we have the observation vector y = _1, z_2, z_3T, the estimator matrix W = _1, w_2, w_3/math> as a row vector, and the estimated variable x = z_4 as a scalar quantity. The autocorrelation matrix C_Y is defined as :C_Y=\left begin E[z_,z_&_E[z_,z_.html"_;"title="_,z_.html"_;"title="begin E[z_,z_">begin E[z_,z_&_E[z_,z_">_,z_.html"_;"title="begin E[z_,z_">begin E[z_,z_&_E[z_,z_&_E[z_,z_.html" ;"title="_,z_">begin E[z_,z_&_E[z_,z_.html" ;"title="_,z_.html" ;"title="begin E[z_,z_">begin E[z_,z_& E[z_,z_">_,z_.html" ;"title="begin E[z_,z_">begin E[z_,z_& E[z_,z_& E[z_,z_">_,z_">begin E[z_,z_&_E[z_,z_.html" ;"title="_,z_.html" ;"title="begin E[z_,z_">begin E[z_,z_& E[z_,z_">_,z_.html" ;"title="begin E[z_,z_">begin E[z_,z_& E[z_,z_& E[z_,z_\ E[z_,z_] & E[z_,z_] & E[z_,z_]\\ E[z_,z_] & E[z_,z_] & E[z_,z_]\end\right]=\left[\begin 1 & 2 & 3\\ 2 & 5 & 8\\ 3 & 8 & 6\end\right]. The cross correlation matrix C_ is defined as :C_=\left begin E[z_,z_\ E[z_,z_.html"_;"title="_,z_.html"_;"title="begin E[z_,z_">begin E[z_,z_\ E[z_,z_">_,z_.html"_;"title="begin E[z_,z_">begin E[z_,z_\ E[z_,z_\ E[z_,z_.html" ;"title="_,z_">begin E[z_,z_\ E[z_,z_.html" ;"title="_,z_.html" ;"title="begin E[z_,z_">begin E[z_,z_\ E[z_,z_">_,z_.html" ;"title="begin E[z_,z_">begin E[z_,z_\ E[z_,z_\ E[z_,z_">_,z_">begin E[z_,z_\ E[z_,z_.html" ;"title="_,z_.html" ;"title="begin E[z_,z_">begin E[z_,z_\ E[z_,z_">_,z_.html" ;"title="begin E[z_,z_">begin E[z_,z_\ E[z_,z_\ E[z_,z_end\right]=\left[\begin 4\\ 9\\ 10\end\right]. We now solve the equation C_Y W^T=C_ by inverting C_Y and pre-multiplying to get :C_Y^C_=\left[\begin 4.85 & -1.71 & -0.142\\ -1.71 & 0.428 & 0.2857\\ -0.142 & 0.2857 & -0.1429\end\right]\left begin 4\\ 9\\ 10\end\right\left begin 2.57\\ -0.142\\ 0.5714\end\rightW^T. So we have w_1=2.57, w_2=-0.142, and w_=.5714 as the optimal coefficients for \hat z_4. Computing the minimum mean square error then gives \left\Vert e\right\Vert _^2=\operatorname _4 z_4WC_=15-WC_=.2857.Moon and Stirling. Note that it is not necessary to obtain an explicit matrix inverse of C_Y to compute the value of W. The matrix equation can be solved by well known methods such as Gauss elimination method. A shorter, non-numerical example can be found in
orthogonality principle In statistics and signal processing, the orthogonality principle is a necessary and sufficient condition for the optimality of a Bayesian estimator. Loosely stated, the orthogonality principle says that the error vector of the optimal estimator (i ...
.


Example 2

Consider a vector y formed by taking N observations of a fixed but unknown scalar parameter x disturbed by white Gaussian noise. We can describe the process by a linear equation y = 1x+ z, where 1 = ,1,\ldots,1T. Depending on context it will be clear if 1 represents a scalar or a vector. Suppose that we know x_0,x_0/math> to be the range within which the value of x is going to fall in. We can model our uncertainty of x by an aprior
uniform distribution Uniform distribution may refer to: * Continuous uniform distribution * Discrete uniform distribution * Uniform distribution (ecology) * Equidistributed sequence See also * * Homogeneous distribution In mathematics, a homogeneous distribution ...
over an interval x_0,x_0/math>, and thus x will have variance of \sigma_X^2 = x_0^2/3.. Let the noise vector z be normally distributed as N(0,\sigma_Z^2I) where I is an identity matrix. Also x and z are independent and C_ = 0. It is easy to see that : \begin & \operatorname\ = 0, \\ & C_Y = \operatorname\ = \sigma_X^2 11^T + \sigma_Z^2I, \\ & C_ = \operatorname\ = \sigma_X^2 1^T. \end Thus, the linear MMSE estimator is given by : \begin \hat &= C_C_Y^ y \\ &= \sigma_X^2 1^T(\sigma_X^2 11^T + \sigma_Z^2I)^ y. \end We can simplify the expression by using the alternative form for W as : \begin \hat &= \left(1^T \fracI 1 + \frac\right)^ 1^T \frac I y \\ &= \frac \left( \frac + \frac \right)^ 1^T y \\ &= \frac \bar, \end where for y = _1,y_2,\ldots,y_NT we have \bar = \frac = \frac. Similarly, the variance of the estimator is :\sigma_^2 = C_C_Y^C_ = \Big(\frac\Big) \sigma_X^2. Thus the MMSE of this linear estimator is :\operatorname = \sigma_X^2 - \sigma_^2 = \Big(\frac\Big) \frac N. For very large N, we see that the MMSE estimator of a scalar with uniform aprior distribution can be approximated by the arithmetic average of all the observed data :\hat = \frac 1 N \sum_^N y_i, while the variance will be unaffected by data \sigma_^2 = \sigma_^2, and the LMMSE of the estimate will tend to zero. However, the estimator is suboptimal since it is constrained to be linear. Had the random variable x also been Gaussian, then the estimator would have been optimal. Notice, that the form of the estimator will remain unchanged, regardless of the apriori distribution of x, so long as the mean and variance of these distributions are the same.


Example 3

Consider a variation of the above example: Two candidates are standing for an election. Let the fraction of votes that a candidate will receive on an election day be x \in ,1 Thus the fraction of votes the other candidate will receive will be 1-x. We shall take x as a random variable with a uniform prior distribution over ,1/math> so that its mean is \bar = 1/2 and variance is \sigma_X^2 = 1/12. A few weeks before the election, two independent public opinion polls were conducted by two different pollsters. The first poll revealed that the candidate is likely to get y_1 fraction of votes. Since some error is always present due to finite sampling and the particular polling methodology adopted, the first pollster declares their estimate to have an error z_1 with zero mean and variance \sigma_^2. Similarly, the second pollster declares their estimate to be y_2 with an error z_2 with zero mean and variance \sigma_^2. Note that except for the mean and variance of the error, the error distribution is unspecified. How should the two polls be combined to obtain the voting prediction for the given candidate? As with previous example, we have : \begin y_1 &= x + z_1 \\ y_2 &= x + z_2. \end Here, both the \operatorname\ = \operatorname\ = \bar = 1/2. Thus, we can obtain the LMMSE estimate as the linear combination of y_1 and y_2 as : \hat = w_1 (y_1 - \bar) + w_2 (y_2 - \bar) + \bar, where the weights are given by : \begin w_1 &= \frac, \\ w_2 &= \frac. \end Here, since the denominator term is constant, the poll with lower error is given higher weight in order to predict the election outcome. Lastly, the variance of \hat is given by : \sigma_^2 = \frac \sigma_X^2 , which makes \sigma_^2 smaller than \sigma_X^2. Thus, the LMMSE is given by :\mathrm = \sigma_^2 - \sigma_^2 = \frac. In general, if we have N pollsters, then \hat = \sum_^N w_i (y_i - \bar) + \bar, where the weight for ''i''-th pollster is given by w_i = \frac and the LMMSE is given by \mathrm = \frac.


Example 4

Suppose that a musician is playing an instrument and that the sound is received by two microphones, each of them located at two different places. Let the attenuation of sound due to distance at each microphone be a_1 and a_2, which are assumed to be known constants. Similarly, let the noise at each microphone be z_1 and z_2, each with zero mean and variances \sigma_^2 and \sigma_^2 respectively. Let x denote the sound produced by the musician, which is a random variable with zero mean and variance \sigma_X^2. How should the recorded music from these two microphones be combined, after being synced with each other? We can model the sound received by each microphone as : \begin y_1 &= a_1 x + z_1 \\ y_2 &= a_2 x + z_2. \end Here both the \operatorname\ = \operatorname\ = 0. Thus, we can combine the two sounds as :y = w_1 y_1 + w_2 y_2 where the ''i''-th weight is given as :w_i = \frac.


See also

*
Bayesian estimator In estimation theory and decision theory, a Bayes estimator or a Bayes action is an estimator or decision rule that minimizes the posterior expected value of a loss function (i.e., the posterior expected loss). Equivalently, it maximizes the po ...
* Mean squared error *
Least squares The method of least squares is a standard approach in regression analysis to approximate the solution of overdetermined systems (sets of equations in which there are more equations than unknowns) by minimizing the sum of the squares of the re ...
* Minimum-variance unbiased estimator (MVUE) *
Orthogonality principle In statistics and signal processing, the orthogonality principle is a necessary and sufficient condition for the optimality of a Bayesian estimator. Loosely stated, the orthogonality principle says that the error vector of the optimal estimator (i ...
* Wiener filter * Kalman filter * Linear prediction * Zero-forcing equalizer


Notes


Further reading

* * * * * * * * * {{DEFAULTSORT:Minimum Mean Square Error Statistical deviation and dispersion Point estimation performance Signal estimation