Neyman–Pearson Lemma
   HOME

TheInfoList



OR:

In
statistics Statistics (from German language, German: ', "description of a State (polity), state, a country") is the discipline that concerns the collection, organization, analysis, interpretation, and presentation of data. In applying statistics to a s ...
, the Neyman–Pearson lemma describes the existence and uniqueness of the likelihood ratio as a uniformly most powerful test in certain contexts. It was introduced by
Jerzy Neyman Jerzy Spława-Neyman (April 16, 1894 – August 5, 1981; ) was a Polish mathematician and statistician who first introduced the modern concept of a confidence interval into statistical hypothesis testing and, with Egon Pearson, revised Ronald Fis ...
and
Egon Pearson Egon Sharpe Pearson (11 August 1895 – 12 June 1980) was one of three children of Karl Pearson and Maria, née Sharpe, and, like his father, a British statistician. Career Pearson was educated at Winchester College and Trinity College ...
in a paper in 1933. The Neyman–Pearson lemma is part of the Neyman–Pearson theory of statistical testing, which introduced concepts such as errors of the second kind,
power function In mathematics, exponentiation, denoted , is an operation involving two numbers: the ''base'', , and the ''exponent'' or ''power'', . When is a positive integer, exponentiation corresponds to repeated multiplication of the base: that is, i ...
, and inductive behavior.The Fisher, Neyman–Pearson Theories of Testing Hypotheses: One Theory or Two?: Journal of the American Statistical Association: Vol 88, No 424
The Fisher, Neyman–Pearson Theories of Testing Hypotheses: One Theory or Two?: Journal of the American Statistical Association: Vol 88, No 424
/ref>Wald: Chapter II: The Neyman–Pearson Theory of Testing a Statistical Hypothesis
Wald: Chapter II: The Neyman–Pearson Theory of Testing a Statistical Hypothesis
/ref>The Empire of Chance
The Empire of Chance
/ref> The previous Fisherian theory of significance testing postulated only one hypothesis. By introducing a competing hypothesis, the Neyman–Pearsonian flavor of statistical testing allows investigating the two types of errors. The trivial cases where one always rejects or accepts the null hypothesis are of little interest but it does prove that one must not relinquish control over one type of error while calibrating the other. Neyman and Pearson accordingly proceeded to restrict their attention to the class of all \alpha level tests while subsequently minimizing type II error, traditionally denoted by \beta. Their seminal paper of 1933, including the Neyman–Pearson lemma, comes at the end of this endeavor, not only showing the existence of tests with the most
power Power may refer to: Common meanings * Power (physics), meaning "rate of doing work" ** Engine power, the power put out by an engine ** Electric power, a type of energy * Power (social and political), the ability to influence people or events Math ...
that retain a prespecified level of type I error (\alpha), but also providing a way to construct such tests. The Karlin-Rubin theorem extends the Neyman–Pearson lemma to settings involving composite hypotheses with monotone likelihood ratios.


Statement

Consider a test with hypotheses H_0: \theta = \theta_0 and H_1:\theta=\theta_1, where the
probability density function In probability theory, a probability density function (PDF), density function, or density of an absolutely continuous random variable, is a Function (mathematics), function whose value at any given sample (or point) in the sample space (the s ...
(or
probability mass function In probability and statistics, a probability mass function (sometimes called ''probability function'' or ''frequency function'') is a function that gives the probability that a discrete random variable is exactly equal to some value. Sometimes i ...
) is \rho(x\mid \theta_i) for i=0,1. For any hypothesis test with rejection set R, and any \alpha\in
, 1 The comma is a punctuation mark that appears in several variants in different languages. Some typefaces render it as a small line, slightly curved or straight, but inclined from the vertical; others give it the appearance of a miniature fille ...
/math>, we say that it satisfies condition P_\alpha if * \alpha = _(X\in R) ** That is, the test has
size Size in general is the Magnitude (mathematics), magnitude or dimensions of a thing. More specifically, ''geometrical size'' (or ''spatial size'') can refer to three geometrical measures: length, area, or volume. Length can be generalized ...
\alpha (that is, the probability of falsely rejecting the null hypothesis is \alpha). * \exists \eta \geq 0 such that :: \begin x\in& R\smallsetminus A\implies \rho(x\mid \theta_1) > \eta \rho(x\mid \theta_0) \\ x\in& R^c\smallsetminus A \implies \rho(x\mid\theta_1) < \eta \rho(x\mid \theta_0) \end : where A is a
negligible set In mathematics, a negligible set is a set that is small enough that it can be ignored for some purpose. As common examples, finite sets can be ignored when studying the limit of a sequence, and null sets can be ignored when studying the integral ...
in both \theta_0 and \theta_1 cases: _(X\in A) = _(X\in A) = 0. * That is, we have a strict likelihood ratio test, except on a negligible subset. For any \alpha\in
, 1 The comma is a punctuation mark that appears in several variants in different languages. Some typefaces render it as a small line, slightly curved or straight, but inclined from the vertical; others give it the appearance of a miniature fille ...
/math>, let the set of level \alpha tests be the set of all hypothesis tests with size at most \alpha. That is, letting its rejection set be R, we have _(X\in R)\leq \alpha. In practice, the likelihood ratio is often used directly to construct tests — see
likelihood-ratio test In statistics, the likelihood-ratio test is a hypothesis test that involves comparing the goodness of fit of two competing statistical models, typically one found by maximization over the entire parameter space and another found after imposing ...
. However it can also be used to suggest particular test-statistics that might be of interest or to suggest simplified tests — for this, one considers algebraic manipulation of the ratio to see if there are key statistics in it related to the size of the ratio (i.e. whether a large statistic corresponds to a small ratio or to a large one).


Example

Let X_1,\dots,X_n be a random sample from the \mathcal(\mu,\sigma^2) distribution where the mean \mu is known, and suppose that we wish to test for H_0:\sigma^2=\sigma_0^2 against H_1:\sigma^2=\sigma_1^2. The likelihood for this set of
normally distributed In probability theory and statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real number, real-valued random variable. The general form of its probability density function is f(x ...
data is :\mathcal\left(\sigma^2\mid\mathbf\right)\propto \left(\sigma^2\right)^ \exp\left\. We can compute the likelihood ratio to find the key statistic in this test and its effect on the test's outcome: :\Lambda(\mathbf) = \frac = \left(\frac\right)^ \exp\left\. This ratio only depends on the data through \sum_^n (x_i-\mu)^2. Therefore, by the Neyman–Pearson lemma, the most powerful test of this type of
hypothesis A hypothesis (: hypotheses) is a proposed explanation for a phenomenon. A scientific hypothesis must be based on observations and make a testable and reproducible prediction about reality, in a process beginning with an educated guess o ...
for this data will depend only on \sum_^n (x_i-\mu)^2. Also, by inspection, we can see that if \sigma_1^2>\sigma_0^2, then \Lambda(\mathbf) is a
decreasing function In mathematics, a monotonic function (or monotone function) is a function between ordered sets that preserves or reverses the given order. This concept first arose in calculus, and was later generalized to the more abstract setting of orde ...
of \sum_^n (x_i-\mu)^2. So we should reject H_0 if \sum_^n (x_i-\mu)^2 is sufficiently large. The rejection threshold depends on the
size Size in general is the Magnitude (mathematics), magnitude or dimensions of a thing. More specifically, ''geometrical size'' (or ''spatial size'') can refer to three geometrical measures: length, area, or volume. Length can be generalized ...
of the test. In this example, the test statistic can be shown to be a scaled chi-square distributed random variable and an exact critical value can be obtained.


Application in economics

A variant of the Neyman–Pearson lemma has found an application in the seemingly unrelated domain of the economics of land value. One of the fundamental problems in
consumer theory The theory of consumer choice is the branch of microeconomics that relates preferences to consumption expenditures and to consumer demand curves. It analyzes how consumers maximize the desirability of their consumption (as measured by their pr ...
is calculating the
demand function In economics, an inverse demand function is the mathematical relationship that expresses price as a function of quantity demanded (it is therefore also known as a price function). Historically, the economists first expressed the price of a good a ...
of the consumer given the prices. In particular, given a heterogeneous land-estate, a price measure over the land, and a subjective utility measure over the land, the consumer's problem is to calculate the best land parcel that they can buy – i.e. the land parcel with the largest utility, whose price is at most their budget. It turns out that this problem is very similar to the problem of finding the most powerful statistical test, and so the Neyman–Pearson lemma can be used.


Uses in electrical engineering

The Neyman–Pearson lemma is quite useful in
electronics engineering Electronic engineering is a sub-discipline of electrical engineering that emerged in the early 20th century and is distinguished by the additional use of active components such as semiconductor devices to amplify and control electric current flow ...
, namely in the design and use of
radar Radar is a system that uses radio waves to determine the distance ('' ranging''), direction ( azimuth and elevation angles), and radial velocity of objects relative to the site. It is a radiodetermination method used to detect and track ...
systems, digital communication systems, and in
signal processing Signal processing is an electrical engineering subfield that focuses on analyzing, modifying and synthesizing ''signals'', such as audio signal processing, sound, image processing, images, Scalar potential, potential fields, Seismic tomograph ...
systems. In radar systems, the Neyman–Pearson lemma is used in first setting the rate of missed detections to a desired (low) level, and then minimizing the rate of
false alarm A false alarm, also called a nuisance alarm, is the deceptive or erroneous report of an emergency, causing unnecessary panic and/or bringing resources (such as emergency services) to a place where they are not needed. False alarms may occur with ...
s, or vice versa. Neither false alarms nor missed detections can be set at arbitrarily low rates, including zero. All of the above goes also for many systems in signal processing.


Uses in particle physics

The Neyman–Pearson lemma is applied to the construction of analysis-specific likelihood-ratios, used to e.g. test for signatures of
new physics Physics beyond the Standard Model (BSM) refers to the theoretical developments needed to explain the deficiencies of the Standard Model, such as the inability to explain the fundamental parameters of the standard model, the strong CP problem, neut ...
against the nominal
Standard Model The Standard Model of particle physics is the Scientific theory, theory describing three of the four known fundamental forces (electromagnetism, electromagnetic, weak interaction, weak and strong interactions – excluding gravity) in the unive ...
prediction in proton–proton collision datasets collected at the
LHC The Large Hadron Collider (LHC) is the world's largest and highest-energy particle accelerator. It was built by the European Organization for Nuclear Research (CERN) between 1998 and 2008, in collaboration with over 10,000 scientists, and ...
.


Discovery of the lemma

Neyman wrote about the discovery of the lemma as follows.Neyman, J. (1970). A glance at some of my personal experiences in the process of research. In ''Scientists at Work: Festschrift in honour of Herman Wold''. Edited by T. Dalenius, G. Karlsson, S. Malmquist. Almqvist & Wiksell, Stockholm. https://worldcat.org/en/title/195948 Paragraph breaks have been inserted.


See also

* Error exponents in hypothesis testing * ''F''-test * Lemma * Wilks' theorem


References

* E. L. Lehmann, Joseph P. Romano, ''Testing statistical hypotheses'', Springer, 2008, p. 60


External links

* Cosma Shalizi gives an intuitive derivation of the Neyman–Pearson Lemm
using ideas from economics

cnx.org: Neyman–Pearson criterion
{{DEFAULTSORT:Neyman-Pearson Lemma Theorems in statistics Statistical tests Articles containing proofs Lemmas