HOME

TheInfoList



OR:

In statistics, the Q-function is the tail distribution function of the
standard normal distribution In statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real-valued random variable. The general form of its probability density function is : f(x) = \frac e^ The parameter \mu i ...
.Basic properties of the Q-function
In other words, Q(x) is the probability that a normal (Gaussian)
random variable A random variable (also called random quantity, aleatory variable, or stochastic variable) is a mathematical formalization of a quantity or object which depends on random events. It is a mapping or a function from possible outcomes (e.g., the p ...
will obtain a value larger than x standard deviations. Equivalently, Q(x) is the probability that a standard normal random variable takes a value larger than x. If Y is a Gaussian random variable with mean \mu and variance \sigma^2, then X = \frac is
standard normal In statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real-valued random variable. The general form of its probability density function is : f(x) = \frac e^ The parameter \mu is ...
and :P(Y > y) = P(X > x) = Q(x) where x = \frac. Other definitions of the ''Q''-function, all of which are simple transformations of the normal
cumulative distribution function In probability theory and statistics, the cumulative distribution function (CDF) of a real-valued random variable X, or just distribution function of X, evaluated at x, is the probability that X will take a value less than or equal to x. Ev ...
, are also used occasionally. Because of its relation to the
cumulative distribution function In probability theory and statistics, the cumulative distribution function (CDF) of a real-valued random variable X, or just distribution function of X, evaluated at x, is the probability that X will take a value less than or equal to x. Ev ...
of the normal distribution, the ''Q''-function can also be expressed in terms of the
error function In mathematics, the error function (also called the Gauss error function), often denoted by , is a complex function of a complex variable defined as: :\operatorname z = \frac\int_0^z e^\,\mathrm dt. This integral is a special (non- elementa ...
, which is an important function in applied mathematics and physics.


Definition and basic properties

Formally, the ''Q''-function is defined as :Q(x) = \frac \int_x^\infty \exp\left(-\frac\right) \, du. Thus, :Q(x) = 1 - Q(-x) = 1 - \Phi(x)\,\!, where \Phi(x) is the cumulative distribution function of the standard normal Gaussian distribution. The ''Q''-function can be expressed in terms of the
error function In mathematics, the error function (also called the Gauss error function), often denoted by , is a complex function of a complex variable defined as: :\operatorname z = \frac\int_0^z e^\,\mathrm dt. This integral is a special (non- elementa ...
, or the complementary error function, as : \begin Q(x) &=\frac\left( \frac \int_^\infty \exp\left(-t^2\right) \, dt \right)\\ &= \frac - \frac \operatorname \left( \frac \right) ~~\text\\ &= \frac\operatorname \left(\frac \right). \end An alternative form of the ''Q''-function known as Craig's formula, after its discoverer, is expressed as: :Q(x) = \frac \int_0^ \exp \left( - \frac \right) d\theta. This expression is valid only for positive values of ''x'', but it can be used in conjunction with ''Q''(''x'') = 1 − ''Q''(−''x'') to obtain ''Q''(''x'') for negative values. This form is advantageous in that the range of integration is fixed and finite. Craig's formula was later extended by Behnad (2020) for the ''Q''-function of the sum of two non-negative variables, as follows: :Q(x+y) = \frac \int_0^ \exp \left( - \frac - \frac \right) d\theta, \quad x,y \geqslant 0 .


Bounds and approximations

*The ''Q''-function is not an
elementary function In mathematics, an elementary function is a function of a single variable (typically real or complex) that is defined as taking sums, products, roots and compositions of finitely many polynomial, rational, trigonometric, hyperbolic, a ...
. However, the Borjesson-Sundberg bounds, where \phi(x) is the density function of the standard normal distribution, ::\left (\frac \right ) \phi(x) < Q(x) < \frac, \qquad x>0, :become increasingly tight for large ''x'', and are often useful. :Using the
substitution Substitution may refer to: Arts and media *Chord substitution, in music, swapping one chord for a related one within a chord progression *Substitution (poetry), a variation in poetic scansion * "Substitution" (song), a 2009 song by Silversun Pic ...
''v'' =''u''2/2, the upper bound is derived as follows: ::Q(x) =\int_x^\infty\phi(u)\,du <\int_x^\infty\frac ux\phi(u)\,du =\int_^\infty\frac\,dv=-\biggl.\frac\biggr, _^\infty=\frac. :Similarly, using \phi'(u) = - u \phi(u) and the
quotient rule In calculus, the quotient rule is a method of finding the derivative of a function that is the ratio of two differentiable functions. Let h(x)=f(x)/g(x), where both and are differentiable and g(x)\neq 0. The quotient rule states that the deriva ...
, ::\left(1+\frac1\right)Q(x) =\int_x^\infty \left(1+\frac1\right)\phi(u)\,du >\int_x^\infty \left(1+\frac1\right)\phi(u)\,du =-\biggl.\fracu\biggr, _x^\infty =\fracx. :Solving for ''Q''(''x'') provides the lower bound. :The
geometric mean In mathematics, the geometric mean is a mean or average which indicates a central tendency of a set of numbers by using the product of their values (as opposed to the arithmetic mean which uses their sum). The geometric mean is defined as the ...
of the upper and lower bound gives a suitable approximation for Q(x): ::Q(x) \approx \frac, \qquad x \geq 0. * Tighter bounds and approximations of Q(x) can also be obtained by optimizing the following expression :: \tilde(x) = \frac. :For x \geq 0, the best upper bound is given by a = 0.344 and b = 5.334 with maximum absolute relative error of 0.44%. Likewise, the best approximation is given by a = 0.339 and b = 5.510 with maximum absolute relative error of 0.27%. Finally, the best lower bound is given by a = 1/\pi and b = 2 \pi with maximum absolute relative error of 1.17%. *The
Chernoff bound In probability theory, the Chernoff bound gives exponentially decreasing bounds on tail distributions of sums of independent random variables. Despite being named after Herman Chernoff, the author of the paper it first appeared in, the result is d ...
of the ''Q''-function is ::Q(x)\leq e^, \qquad x>0 *Improved exponential bounds and a pure exponential approximation are ::Q(x)\leq \tfrace^+\tfrace^ \leq \tfrace^, \qquad x>0 :: Q(x)\approx \frace^+\frace^, \qquad x>0 *The above were generalized by Tanash & Riihonen (2020), who showed that Q(x) can be accurately approximated or bounded by ::\tilde(x) = \sum_^N a_n e^. :In particular, they presented a systematic methodology to solve the numerical coefficients \_^N that yield a
minimax Minimax (sometimes MinMax, MM or saddle point) is a decision rule used in artificial intelligence, decision theory, game theory, statistics, and philosophy for ''mini''mizing the possible loss for a worst case (''max''imum loss) scenario. Whe ...
approximation or bound: Q(x) \approx \tilde(x), Q(x) \leq \tilde(x), or Q(x) \geq \tilde(x) for x\geq0. With the example coefficients tabulated in the paper for N = 20, the relative and absolute approximation errors are less than 2.831 \cdot 10^ and 1.416 \cdot 10^, respectively. The coefficients \_^N for many variations of the exponential approximations and bounds up to N = 25 have been released to open access as a comprehensive dataset. *Another approximation of Q(x) for x \in [0,\infty) is given by Karagiannidis & Lioumpas (2007) who showed for the appropriate choice of parameters \ that :: f(x; A, B) = \frac \approx \operatorname \left(x\right). : The absolute error between f(x; A, B) and \operatorname(x) over the range [0, R] is minimized by evaluating :: \ = \underset \frac \int_0^R , f(x; A, B) - \operatorname(x) , dx. : Using R = 20 and numerically integrating, they found the minimum error occurred when \ = \, which gave a good approximation for \forall x \ge 0. : Substituting these values and using the relationship between Q(x) and \operatorname(x) from above gives :: Q(x)\approx\frac, x \ge 0. : Alternative coefficients are also available for the above 'Karagiannidis–Lioumpas approximation' for tailoring accuracy for a specific application or transforming it into a tight bound. *A tighter and more tractable approximation of Q(x) for positive arguments x \in [0,\infty) is given by López-Benítez & Casadevall (2011) based on a second-order exponential function: :: Q(x) \approx e^, \qquad x \ge 0. : The fitting coefficients (a,b,c) can be optimized over any desired range of arguments in order to minimize the sum of square errors (a = 0.3842, b = 0.7640, c = 0.6964 for x \in [0,20]) or minimize the maximum absolute error (a = 0.4920, b = 0.2887, c = 1.1893 for x \in [0,20]). This approximation offers some benefits such as a good trade-off between accuracy and analytical tractability (for example, the extension to any arbitrary power of Q(x) is trivial and does not alter the algebraic form of the approximation).


Inverse ''Q''

The inverse ''Q''-function can be related to the inverse error functions: :Q^(y) = \sqrt\ \mathrm^(1-2y) = \sqrt\ \mathrm^(2y) The function Q^(y) finds application in digital communications. It is usually expressed in dB and generally called Q-factor: :\mathrm = 20 \log_\!\left(Q^(y)\right)\!~\mathrm where ''y'' is the bit-error rate (BER) of the digitally modulated signal under analysis. For instance, for
QPSK Phase-shift keying (PSK) is a digital modulation process which conveys data by changing (modulating) the phase of a constant frequency reference signal (the carrier wave). The modulation is accomplished by varying the sine and cosine inputs at ...
in additive white Gaussian noise, the Q-factor defined above coincides with the value in dB of the
signal to noise ratio In signal processing, a signal is a function that conveys information about a phenomenon. Any quantity that can vary over space or time can be used as a signal to share messages between observers. The ''IEEE Transactions on Signal Processing'' ...
that yields a bit error rate equal to ''y''.


Values

The ''Q''-function is well tabulated and can be computed directly in most of the mathematical software packages such as R and those available in Python,
MATLAB MATLAB (an abbreviation of "MATrix LABoratory") is a proprietary multi-paradigm programming language and numeric computing environment developed by MathWorks. MATLAB allows matrix manipulations, plotting of functions and data, implementa ...
and
Mathematica Wolfram Mathematica is a software system with built-in libraries for several areas of technical computing that allow machine learning, statistics, symbolic computation, data manipulation, network analysis, time series analysis, NLP, optimi ...
. Some values of the ''Q''-function are given below for reference.


Generalization to high dimensions

The ''Q''-function can be generalized to higher dimensions: :Q(\mathbf)= \mathbb(\mathbf\geq \mathbf), where \mathbf\sim \mathcal(\mathbf,\, \Sigma) follows the multivariate normal distribution with covariance \Sigma and the threshold is of the form \mathbf=\gamma\Sigma\mathbf^* for some positive vector \mathbf^*>\mathbf and positive constant \gamma>0. As in the one dimensional case, there is no simple analytical formula for the ''Q''-function. Nevertheless, the ''Q''-function can b
approximated arbitrarily well
as \gamma becomes larger and larger.{{cite book , chapter=Logarithmically efficient estimation of the tail of the multivariate normal distribution , last1=Botev , first1=Z. I. , last2=Mackinlay , first2=D. , last3=Chen , first3=Y.-L. , date=2017 , publisher=IEEE , isbn=978-1-5386-3428-8 , title= 2017 Winter Simulation Conference (WSC), pages=1903–191 , doi= 10.1109/WSC.2017.8247926 , s2cid=4626481


References

Normal distribution Special functions Functions related to probability distributions Articles containing proofs