HOME

TheInfoList



OR:

In
statistical hypothesis testing A statistical hypothesis test is a method of statistical inference used to decide whether the data at hand sufficiently support a particular hypothesis. Hypothesis testing allows us to make probabilistic statements about population parameters. ...
, the error exponent of a hypothesis testing procedure is the rate at which the probabilities of Type I and Type II decay exponentially with the size of the sample used in the test. For example, if the probability of error P_ of a test decays as e^, where n is the sample size, the error exponent is \beta. Formally, the error exponent of a test is defined as the limiting value of the ratio of the negative logarithm of the error probability to the sample size for large sample sizes: \lim_\frac. Error exponents for different hypothesis tests are computed using
Sanov's theorem In mathematics and information theory, Sanov's theorem gives a bound on the probability of observing an atypical sequence of samples from a given probability distribution. In the language of large deviations theory, Sanov's theorem identifies t ...
and other results from
large deviations theory In probability theory, the theory of large deviations concerns the asymptotic behaviour of remote tails of sequences of probability distributions. While some basic ideas of the theory can be traced to Laplace, the formalization started with insura ...
.


Error exponents in binary hypothesis testing

Consider a binary hypothesis testing problem in which observations are modeled as
independent and identically distributed random variables In probability theory and statistics, a collection of random variables is independent and identically distributed if each random variable has the same probability distribution as the others and all are mutually independent. This property is us ...
under each hypothesis. Let Y_1, Y_2, \ldots, Y_n denote the observations. Let f_0 denote the
probability density function In probability theory, a probability density function (PDF), or density of a continuous random variable, is a function whose value at any given sample (or point) in the sample space (the set of possible values taken by the random variable) can ...
of each observation Y_i under the null hypothesis H_0 and let f_1 denote the probability density function of each observation Y_i under the alternate hypothesis H_1. In this case there are two possible error events. Error of type 1, also called
false positive A false positive is an error in binary classification in which a test result incorrectly indicates the presence of a condition (such as a disease when the disease is not present), while a false negative is the opposite error, where the test result ...
, occurs when the null hypothesis is true and it is wrongly rejected. Error of type 2, also called false negative, occurs when the alternate hypothesis is true and null hypothesis is not rejected. The probability of type 1 error is denoted P (\mathrm\mid H_0) and the probability of type 2 error is denoted P (\mathrm\mid H_1).


Optimal error exponent for Neyman–Pearson testing

In the Neyman–Pearson version of binary hypothesis testing, one is interested in minimizing the probability of type 2 error P (\text\mid H_1) subject to the constraint that the probability of type 1 error P (\text\mid H_0) is less than or equal to a pre-specified level \alpha. In this setting, the optimal testing procedure is a
likelihood-ratio test In statistics, the likelihood-ratio test assesses the goodness of fit of two competing statistical models based on the ratio of their likelihoods, specifically one found by maximization over the entire parameter space and another found after im ...
. Furthermore, the optimal test guarantees that the type 2 error probability decays exponentially in the sample size n according to \lim_ \frac = D(f_0\parallel f_1). The error exponent D(f_0\parallel f_1) is the
Kullback–Leibler divergence In mathematical statistics, the Kullback–Leibler divergence (also called relative entropy and I-divergence), denoted D_\text(P \parallel Q), is a type of statistical distance: a measure of how one probability distribution ''P'' is different fro ...
between the probability distributions of the observations under the two hypotheses. This exponent is also referred to as the Chernoff–Stein lemma exponent.


Optimal error exponent for average error probability in Bayesian hypothesis testing

In the
Bayesian Thomas Bayes (/beɪz/; c. 1701 – 1761) was an English statistician, philosopher, and Presbyterian minister. Bayesian () refers either to a range of concepts and approaches that relate to statistical methods based on Bayes' theorem, or a followe ...
version of binary hypothesis testing one is interested in minimizing the average error probability under both hypothesis, assuming a prior probability of occurrence on each hypothesis. Let \pi_0 denote the prior probability of hypothesis H_0 . In this case the average error probability is given by P_\text = \pi_0 P (\text\mid H_0) + (1-\pi_0)P (\text\mid H_1). In this setting again a likelihood ratio test is optimal and the optimal error decays as \lim_ \frac = C(f_0,f_1) where C(f_0,f_1) represents the Chernoff-information between the two distributions defined as C(f_0,f_1) = \max_ \left \ln \int (f_0(x))^\lambda (f_1(x))^ \, dx \right/math>.


References

{{Reflist Statistical hypothesis testing Information theory Large deviations theory