In
statistics
Statistics (from German language, German: ', "description of a State (polity), state, a country") is the discipline that concerns the collection, organization, analysis, interpretation, and presentation of data. In applying statistics to a s ...
a minimum-variance unbiased estimator (MVUE) or uniformly minimum-variance unbiased estimator (UMVUE) is an
unbiased estimator
In statistics, the bias of an estimator (or bias function) is the difference between this estimator's expected value and the true value of the parameter being estimated. An estimator or decision rule with zero bias is called ''unbiased''. In stat ...
that has lower variance than any other unbiased estimator for all possible values of the parameter.
For practical statistics problems, it is important to determine the MVUE if one exists, since less-than-optimal procedures would naturally be avoided, other things being equal. This has led to substantial development of statistical theory related to the problem of optimal estimation.
While combining the constraint of
unbiasedness
In statistics, the bias of an estimator (or bias function) is the difference between this estimator's expected value and the true value of the parameter being estimated. An estimator or decision rule with zero bias is called ''unbiased''. In stat ...
with the desirability metric of least
variance
In probability theory and statistics, variance is the expected value of the squared deviation from the mean of a random variable. The standard deviation (SD) is obtained as the square root of the variance. Variance is a measure of dispersion ...
leads to good results in most practical settings—making MVUE a natural starting point for a broad range of analyses—a targeted specification may perform better for a given problem; thus, MVUE is not always the best stopping point.
Definition
Consider estimation of
based on data
i.i.d. from some member of a family of densities
, where
is the parameter space. An unbiased estimator
of
is ''UMVUE'' if
,
:
for any other unbiased estimator
If an unbiased estimator of
exists, then one can prove there is an essentially unique MVUE.
Using the
Rao–Blackwell theorem one can also prove that determining the MVUE is simply a matter of finding a
complete sufficient statistic for the family
and conditioning ''any'' unbiased estimator on it.
Further, by the
Lehmann–Scheffé theorem, an unbiased estimator that is a function of a complete, sufficient statistic is the UMVUE estimator.
Put formally, suppose
is unbiased for
, and that
is a complete sufficient statistic for the family of densities. Then
:
is the MVUE for
A
Bayesian analog is a
Bayes estimator
In estimation theory and decision theory, a Bayes estimator or a Bayes action is an estimator or decision rule that minimizes the posterior expected value of a loss function (i.e., the posterior expected loss). Equivalently, it maximizes the ...
, particularly with
minimum mean square error
In statistics and signal processing, a minimum mean square error (MMSE) estimator is an estimation method which minimizes the mean square error (MSE), which is a common measure of estimator quality, of the fitted values of a dependent variable. I ...
(MMSE).
Estimator selection
An
efficient estimator
In statistics, efficiency is a measure of quality of an estimator, of an experimental design, or of a hypothesis testing procedure. Essentially, a more efficient estimator needs fewer input data or observations than a less efficient one to achiev ...
need not exist, but if it does and if it is unbiased,
it is the MVUE. Since the
mean squared error
In statistics, the mean squared error (MSE) or mean squared deviation (MSD) of an estimator (of a procedure for estimating an unobserved quantity) measures the average of the squares of the errors—that is, the average squared difference betwee ...
(MSE) of an estimator ''δ'' is
:
the MVUE minimizes MSE ''among unbiased estimators''. In some cases biased estimators have lower MSE because they have a smaller variance than does any unbiased estimator; see
estimator bias
In statistics, the bias of an estimator (or bias function) is the difference between this estimator's expected value and the true value of the parameter being estimated. An estimator or decision rule with zero bias is called ''unbiased''. In stat ...
.
Example
Consider the data to be a single observation from an
absolutely continuous distribution on
with density
:
where ''θ > 0'', and we wish to find the UMVU estimator of
:
First we recognize that the density can be written as
:
which is an exponential family with
sufficient statistic
In statistics, sufficiency is a property of a statistic computed on a sample dataset in relation to a parametric model of the dataset. A sufficient statistic contains all of the information that the dataset provides about the model parameters. It ...
. In fact this is a full rank exponential family, and therefore
is complete sufficient. See
exponential family
In probability and statistics, an exponential family is a parametric set of probability distributions of a certain form, specified below. This special form is chosen for mathematical convenience, including the enabling of the user to calculate ...
for a derivation which shows
:
Therefore,
:
Here we use Lehmann–Scheffé theorem to get the MVUE. Clearly,
is unbiased and
is complete sufficient, thus the UMVU estimator is
:
This example illustrates that an unbiased function of the complete sufficient statistic will be UMVU, as
Lehmann–Scheffé theorem states.
Other examples
* For a normal distribution with unknown mean and variance, the
sample mean
The sample mean (sample average) or empirical mean (empirical average), and the sample covariance or empirical covariance are statistics computed from a sample of data on one or more random variables.
The sample mean is the average value (or me ...
and (unbiased)
sample variance
In probability theory and statistics, variance is the expected value of the squared deviation from the mean of a random variable. The standard deviation (SD) is obtained as the square root of the variance. Variance is a measure of dispersion, ...
are the MVUEs for the population mean and population variance.
*:However, the
sample standard deviation
In statistics, the standard deviation is a measure of the amount of variation of the values of a variable about its mean. A low standard deviation indicates that the values tend to be close to the mean (also called the expected value) of the ...
is not unbiased for the population standard deviation – see
unbiased estimation of standard deviation.
*:Further, for other distributions the sample mean and sample variance are not in general MVUEs – for a
uniform distribution with unknown upper and lower bounds, the
mid-range
In statistics, the mid-range or mid-extreme is a measure of central tendency of a sample defined as the arithmetic mean of the maximum and minimum values of the data set:
:M=\frac.
The mid-range is closely related to the range, a measure of ...
is the MVUE for the population mean.
* If ''k'' exemplars are chosen (without replacement) from a
discrete uniform distribution
In probability theory and statistics, the discrete uniform distribution is a symmetric probability distribution wherein each of some finite whole number ''n'' of outcome values are equally likely to be observed. Thus every one of the ''n'' out ...
over the set with unknown upper bound ''N'', the MVUE for ''N'' is:
::
:where ''m'' is the
sample maximum
In statistics, the sample maximum and sample minimum, also called the largest observation and smallest observation, are the values of the greatest and least elements of a sample. They are basic summary statistics, used in descriptive statistics ...
. This is a scaled and shifted (so unbiased) transform of the sample maximum, which is a sufficient and complete statistic. See
German tank problem
German(s) may refer to:
* Germany, the country of the Germans and German things
**Germania (Roman era)
* Germans, citizens of Germany, people of German ancestry, or native speakers of the German language
** For citizenship in Germany, see also Ge ...
for details.
See also
*
Cramér–Rao bound
In estimation theory and statistics, the Cramér–Rao bound (CRB) relates to estimation of a deterministic (fixed, though unknown) parameter. The result is named in honor of Harald Cramér and Calyampudi Radhakrishna Rao, but has also been d ...
*
Best linear unbiased estimator
Best or The Best may refer to:
People
* Best (surname), people with the surname Best
* Best (footballer, born 1968), retired Portuguese footballer
Companies and organizations
* Best & Co., an 1879–1971 clothing chain
* Best Lock Corporatio ...
(BLUE)
*
Bias–variance tradeoff
In statistics and machine learning, the bias–variance tradeoff describes the relationship between a model's complexity, the accuracy of its predictions, and how well it can make predictions on previously unseen data that were not used to train ...
*
Lehmann–Scheffé theorem
*
U-statistic
In statistical theory, a U-statistic is a class of statistics defined as the average over the application of a given function applied to all tuples of a fixed size. The letter "U" stands for unbiased. In elementary statistics, U-statistics arise ...
Bayesian analogs
*
Bayes estimator
In estimation theory and decision theory, a Bayes estimator or a Bayes action is an estimator or decision rule that minimizes the posterior expected value of a loss function (i.e., the posterior expected loss). Equivalently, it maximizes the ...
*
Minimum mean square error
In statistics and signal processing, a minimum mean square error (MMSE) estimator is an estimation method which minimizes the mean square error (MSE), which is a common measure of estimator quality, of the fitted values of a dependent variable. I ...
(MMSE)
References
*
* Keener, Robert W. (2010). ''Theoretical statistics: Topics for a core course''. New York: Springer.
DOI 10.1007/978-0-387-93839-4
*
{{Statistics, inference, collapsed
Estimator