HOME

TheInfoList



OR:

In
statistics Statistics (from German language, German: ', "description of a State (polity), state, a country") is the discipline that concerns the collection, organization, analysis, interpretation, and presentation of data. In applying statistics to a s ...
a minimum-variance unbiased estimator (MVUE) or uniformly minimum-variance unbiased estimator (UMVUE) is an
unbiased estimator In statistics, the bias of an estimator (or bias function) is the difference between this estimator's expected value and the true value of the parameter being estimated. An estimator or decision rule with zero bias is called ''unbiased''. In stat ...
that has lower variance than any other unbiased estimator for all possible values of the parameter. For practical statistics problems, it is important to determine the MVUE if one exists, since less-than-optimal procedures would naturally be avoided, other things being equal. This has led to substantial development of statistical theory related to the problem of optimal estimation. While combining the constraint of
unbiasedness In statistics, the bias of an estimator (or bias function) is the difference between this estimator's expected value and the true value of the parameter being estimated. An estimator or decision rule with zero bias is called ''unbiased''. In stat ...
with the desirability metric of least
variance In probability theory and statistics, variance is the expected value of the squared deviation from the mean of a random variable. The standard deviation (SD) is obtained as the square root of the variance. Variance is a measure of dispersion ...
leads to good results in most practical settings—making MVUE a natural starting point for a broad range of analyses—a targeted specification may perform better for a given problem; thus, MVUE is not always the best stopping point.


Definition

Consider estimation of g(\theta) based on data X_1, X_2, \ldots, X_n i.i.d. from some member of a family of densities p_\theta, \theta \in \Omega, where \Omega is the parameter space. An unbiased estimator \delta(X_1, X_2, \ldots, X_n) of g(\theta) is ''UMVUE'' if \forall \theta \in \Theta, : \operatorname(\delta(X_1, X_2, \ldots, X_n)) \leq \operatorname(\tilde(X_1, X_2, \ldots, X_n)) for any other unbiased estimator \tilde. If an unbiased estimator of g(\theta) exists, then one can prove there is an essentially unique MVUE. Using the Rao–Blackwell theorem one can also prove that determining the MVUE is simply a matter of finding a complete sufficient statistic for the family p_\theta, \theta \in \Omega and conditioning ''any'' unbiased estimator on it. Further, by the Lehmann–Scheffé theorem, an unbiased estimator that is a function of a complete, sufficient statistic is the UMVUE estimator. Put formally, suppose \delta(X_1, X_2, \ldots, X_n) is unbiased for g(\theta), and that T is a complete sufficient statistic for the family of densities. Then : \eta(X_1, X_2, \ldots, X_n) = \operatorname(\delta(X_1, X_2, \ldots, X_n)\mid T)\, is the MVUE for g(\theta). A Bayesian analog is a
Bayes estimator In estimation theory and decision theory, a Bayes estimator or a Bayes action is an estimator or decision rule that minimizes the posterior expected value of a loss function (i.e., the posterior expected loss). Equivalently, it maximizes the ...
, particularly with
minimum mean square error In statistics and signal processing, a minimum mean square error (MMSE) estimator is an estimation method which minimizes the mean square error (MSE), which is a common measure of estimator quality, of the fitted values of a dependent variable. I ...
(MMSE).


Estimator selection

An
efficient estimator In statistics, efficiency is a measure of quality of an estimator, of an experimental design, or of a hypothesis testing procedure. Essentially, a more efficient estimator needs fewer input data or observations than a less efficient one to achiev ...
need not exist, but if it does and if it is unbiased, it is the MVUE. Since the
mean squared error In statistics, the mean squared error (MSE) or mean squared deviation (MSD) of an estimator (of a procedure for estimating an unobserved quantity) measures the average of the squares of the errors—that is, the average squared difference betwee ...
(MSE) of an estimator ''δ'' is : \operatorname(\delta) = \operatorname(\delta) + \operatorname(\delta)2 \ the MVUE minimizes MSE ''among unbiased estimators''. In some cases biased estimators have lower MSE because they have a smaller variance than does any unbiased estimator; see
estimator bias In statistics, the bias of an estimator (or bias function) is the difference between this estimator's expected value and the true value of the parameter being estimated. An estimator or decision rule with zero bias is called ''unbiased''. In stat ...
.


Example

Consider the data to be a single observation from an absolutely continuous distribution on \mathbb with density : p_\theta(x) = \frac, where ''θ > 0'', and we wish to find the UMVU estimator of : g(\theta) = \frac 1 First we recognize that the density can be written as : \frac \exp( -\theta \log(1 + e^) + \log(\theta)) which is an exponential family with
sufficient statistic In statistics, sufficiency is a property of a statistic computed on a sample dataset in relation to a parametric model of the dataset. A sufficient statistic contains all of the information that the dataset provides about the model parameters. It ...
T = \log(1 + e^). In fact this is a full rank exponential family, and therefore T is complete sufficient. See
exponential family In probability and statistics, an exponential family is a parametric set of probability distributions of a certain form, specified below. This special form is chosen for mathematical convenience, including the enabling of the user to calculate ...
for a derivation which shows : \operatorname(T) = \frac 1 \theta,\quad \operatorname(T) = \frac 1 Therefore, : \operatorname(T^2) = \frac 2 Here we use Lehmann–Scheffé theorem to get the MVUE. Clearly, \delta(X) = T^2/2 is unbiased and T = \log(1 + e^) is complete sufficient, thus the UMVU estimator is : \eta(X) = \operatorname(\delta(X) \mid T) = \operatorname \left( \left. \frac 2 \,\\, T \right) = \frac 2 = \frac 2 This example illustrates that an unbiased function of the complete sufficient statistic will be UMVU, as Lehmann–Scheffé theorem states.


Other examples

* For a normal distribution with unknown mean and variance, the
sample mean The sample mean (sample average) or empirical mean (empirical average), and the sample covariance or empirical covariance are statistics computed from a sample of data on one or more random variables. The sample mean is the average value (or me ...
and (unbiased)
sample variance In probability theory and statistics, variance is the expected value of the squared deviation from the mean of a random variable. The standard deviation (SD) is obtained as the square root of the variance. Variance is a measure of dispersion, ...
are the MVUEs for the population mean and population variance. *:However, the
sample standard deviation In statistics, the standard deviation is a measure of the amount of variation of the values of a variable about its mean. A low standard deviation indicates that the values tend to be close to the mean (also called the expected value) of the ...
is not unbiased for the population standard deviation – see unbiased estimation of standard deviation. *:Further, for other distributions the sample mean and sample variance are not in general MVUEs – for a uniform distribution with unknown upper and lower bounds, the
mid-range In statistics, the mid-range or mid-extreme is a measure of central tendency of a sample defined as the arithmetic mean of the maximum and minimum values of the data set: :M=\frac. The mid-range is closely related to the range, a measure of ...
is the MVUE for the population mean. * If ''k'' exemplars are chosen (without replacement) from a
discrete uniform distribution In probability theory and statistics, the discrete uniform distribution is a symmetric probability distribution wherein each of some finite whole number ''n'' of outcome values are equally likely to be observed. Thus every one of the ''n'' out ...
over the set with unknown upper bound ''N'', the MVUE for ''N'' is: :: \frac m - 1, :where ''m'' is the
sample maximum In statistics, the sample maximum and sample minimum, also called the largest observation and smallest observation, are the values of the greatest and least elements of a sample. They are basic summary statistics, used in descriptive statistics ...
. This is a scaled and shifted (so unbiased) transform of the sample maximum, which is a sufficient and complete statistic. See
German tank problem German(s) may refer to: * Germany, the country of the Germans and German things **Germania (Roman era) * Germans, citizens of Germany, people of German ancestry, or native speakers of the German language ** For citizenship in Germany, see also Ge ...
for details.


See also

*
Cramér–Rao bound In estimation theory and statistics, the Cramér–Rao bound (CRB) relates to estimation of a deterministic (fixed, though unknown) parameter. The result is named in honor of Harald Cramér and Calyampudi Radhakrishna Rao, but has also been d ...
*
Best linear unbiased estimator Best or The Best may refer to: People * Best (surname), people with the surname Best * Best (footballer, born 1968), retired Portuguese footballer Companies and organizations * Best & Co., an 1879–1971 clothing chain * Best Lock Corporatio ...
(BLUE) *
Bias–variance tradeoff In statistics and machine learning, the bias–variance tradeoff describes the relationship between a model's complexity, the accuracy of its predictions, and how well it can make predictions on previously unseen data that were not used to train ...
* Lehmann–Scheffé theorem *
U-statistic In statistical theory, a U-statistic is a class of statistics defined as the average over the application of a given function applied to all tuples of a fixed size. The letter "U" stands for unbiased. In elementary statistics, U-statistics arise ...


Bayesian analogs

*
Bayes estimator In estimation theory and decision theory, a Bayes estimator or a Bayes action is an estimator or decision rule that minimizes the posterior expected value of a loss function (i.e., the posterior expected loss). Equivalently, it maximizes the ...
*
Minimum mean square error In statistics and signal processing, a minimum mean square error (MMSE) estimator is an estimation method which minimizes the mean square error (MSE), which is a common measure of estimator quality, of the fitted values of a dependent variable. I ...
(MMSE)


References

* * Keener, Robert W. (2010). ''Theoretical statistics: Topics for a core course''. New York: Springer. DOI 10.1007/978-0-387-93839-4 * {{Statistics, inference, collapsed Estimator