HOME

TheInfoList



OR:

In
statistics Statistics (from German language, German: ', "description of a State (polity), state, a country") is the discipline that concerns the collection, organization, analysis, interpretation, and presentation of data. In applying statistics to a s ...
, the Lehmann–Scheffé theorem ties together completeness, sufficiency, uniqueness, and best unbiased estimation. The theorem states that any
estimator In statistics, an estimator is a rule for calculating an estimate of a given quantity based on Sample (statistics), observed data: thus the rule (the estimator), the quantity of interest (the estimand) and its result (the estimate) are distinguish ...
that is
unbiased Bias is a disproportionate weight ''in favor of'' or ''against'' an idea or thing, usually in a way that is inaccurate, closed-minded, prejudicial, or unfair. Biases can be innate or learned. People may develop biases for or against an individ ...
for a given unknown quantity and that depends on the data only through a
complete Complete may refer to: Logic * Completeness (logic) * Completeness of a theory, the property of a theory that every formula in the theory's language or its negation is provable Mathematics * The completeness of the real numbers, which implies t ...
,
sufficient statistic In statistics, sufficiency is a property of a statistic computed on a sample dataset in relation to a parametric model of the dataset. A sufficient statistic contains all of the information that the dataset provides about the model parameters. It ...
is the unique best unbiased estimator of that quantity. The Lehmann–Scheffé theorem is named after
Erich Leo Lehmann Erich Leo Lehmann (20 November 1917 – 12 September 2009) was a German-born American statistician, who made a major contribution to nonparametric hypothesis testing. He is one of the eponyms of the Lehmann–Scheffé theorem and of the Hodges ...
and
Henry Scheffé Henry Scheffé (April 11, 1907 – July 5, 1977) was an American statistician. He is known for the Lehmann–Scheffé theorem and Scheffé's method. Education and career Scheffé was born in New York City on April 11, 1907, the child of Germ ...
, given their two early papers. If T is a complete sufficient statistic for \theta and \operatorname (T)\tau(\theta) then g(T) is the uniformly minimum-variance unbiased estimator (UMVUE) of \tau(\theta).


Statement

Let \vec= X_1, X_2, \dots, X_n be a random sample from a distribution that has p.d.f (or p.m.f in the discrete case) f(x:\theta) where \theta \in \Omega is a parameter in the parameter space. Suppose Y = u(\vec) is a sufficient statistic for ''θ'', and let \ be a complete family. If \varphi:\operatorname
varphi(Y) Phi ( ; uppercase Φ, lowercase φ or ϕ; ''pheî'' ; Modern Greek: ''fi'' ) is the twenty-first letter of the Greek alphabet. In Archaic and Classical Greek (c. 9th to 4th century BC), it represented an aspirated voiceless bilabial plos ...
= \theta then \varphi(Y) is the unique MVUE of ''θ''.


Proof

By the Rao–Blackwell theorem, if Z is an unbiased estimator of ''θ'' then \varphi(Y):= \operatorname \mid Y/math> defines an unbiased estimator of ''θ'' with the property that its variance is not greater than that of Z. Now we show that this function is unique. Suppose W is another candidate MVUE estimator of ''θ''. Then again \psi(Y):= \operatorname \mid Y/math> defines an unbiased estimator of ''θ'' with the property that its variance is not greater than that of W. Then : \operatorname varphi(Y) - \psi(Y)= 0, \theta \in \Omega. Since \ is a complete family : \operatorname varphi(Y) - \psi(Y)= 0 \implies \varphi(y) - \psi(y) = 0, \theta \in \Omega and therefore the function \varphi is the unique function of Y with variance not greater than that of any other unbiased estimator. We conclude that \varphi(Y) is the MVUE.


Example for when using a non-complete minimal sufficient statistic

An example of an improvable Rao–Blackwell improvement, when using a minimal sufficient statistic that is not complete, was provided by Galili and Meilijson in 2016. Let X_1, \ldots, X_n be a random sample from a scale-uniform distribution X \sim U ( (1-k) \theta, (1+k) \theta), with unknown mean \operatorname \theta and known design parameter k \in (0,1). In the search for "best" possible unbiased estimators for \theta, it is natural to consider X_1 as an initial (crude) unbiased estimator for \theta and then try to improve it. Since X_1 is not a function of T = \left( X_, X_ \right), the minimal sufficient statistic for \theta (where X_ = \min_i X_i and X_ = \max_i X_i ), it may be improved using the Rao–Blackwell theorem as follows: :\hat_ =\operatorname_\theta _1\mid X_, X_= \frac 2. However, the following unbiased estimator can be shown to have lower variance: :\hat_ = \frac 1 \cdot \frac 2. And in fact, it could be even further improved when using the following estimator: :\hat_\text=\frac n \left - \frac \right\frac The model is a
scale model A scale model is a physical model that is geometrically similar to an object (known as the ''prototype''). Scale models are generally smaller than large prototypes such as vehicles, buildings, or people; but may be larger than small protot ...
. Optimal equivariant estimators can then be derived for
loss function In mathematical optimization and decision theory, a loss function or cost function (sometimes also called an error function) is a function that maps an event or values of one or more variables onto a real number intuitively representing some "cost ...
s that are invariant.


See also

* Basu's theorem *
Completeness (statistics) In statistics, completeness is a property of a statistic computed on a sample dataset in relation to a parametric model of the dataset. It is opposed to the concept of an ancillary statistic. While an ancillary statistic contains no information ...
* Rao–Blackwell theorem


References

{{DEFAULTSORT:Lehmann-Scheffe theorem Theorems in statistics Estimation theory