Inverse Normal Distribution
   HOME

TheInfoList



OR:

In probability theory, the inverse Gaussian distribution (also known as the Wald distribution) is a two-parameter family of continuous probability distributions with
support Support may refer to: Arts, entertainment, and media * Supporting character Business and finance * Support (technical analysis) * Child support * Customer support * Income Support Construction * Support (structure), or lateral support, a ...
on (0,∞). Its probability density function is given by : f(x;\mu,\lambda) = \sqrt\frac \exp\biggl(-\frac\biggr) for ''x'' > 0, where \mu > 0 is the mean and \lambda > 0 is the shape parameter. The inverse Gaussian distribution has several properties analogous to a Gaussian distribution. The name can be misleading: it is an "inverse" only in that, while the Gaussian describes a Brownian motion's level at a fixed time, the inverse Gaussian describes the distribution of the time a Brownian motion with positive drift takes to reach a fixed positive level. Its cumulant generating function (logarithm of the characteristic function) is the inverse of the cumulant generating function of a Gaussian random variable. To indicate that a
random variable A random variable (also called random quantity, aleatory variable, or stochastic variable) is a mathematical formalization of a quantity or object which depends on random events. It is a mapping or a function from possible outcomes (e.g., the po ...
''X'' is inverse Gaussian-distributed with mean μ and shape parameter λ we write X \sim \operatorname(\mu, \lambda)\,\!.


Properties


Single parameter form

The probability density function (pdf) of the inverse Gaussian distribution has a single parameter form given by : f(x;\mu,\mu^2) = \frac \exp\biggl(-\frac\biggr). In this form, the mean and variance of the distribution are equal, \mathbb = \text(X). Also, the cumulative distribution function (cdf) of the single parameter inverse Gaussian distribution is related to the standard normal distribution by : \begin \Pr(X < x) &= \Phi(z_1) + e^ \Phi(z_2), & \text & \quad 0 < x \leq \mu, \\ \Pr(X > x) &= \Phi(-z_1) - e^ \Phi(z_2), & \text & \quad x \geq \mu. \end where z_1 = \frac - x^, z_2 = \frac + x^, and the \Phi is the cdf of standard normal distribution. The variables z_1 and z_2 are related to each other by the identity z_2^2 = z_1^2 + 4\mu. In the single parameter form, the MGF simplifies to : M(t) = \exp mu(1-\sqrt) An inverse Gaussian distribution in double parameter form f(x;\mu,\lambda) can be transformed into a single parameter form f(y;\mu_0,\mu_0^2) by appropriate scaling y = \frac, where \mu_0 = \mu^3/\lambda. The standard form of inverse Gaussian distribution is : f(x;1,1) = \frac \exp\biggl(-\frac\biggr).


Summation

If ''X''''i'' has an \operatorname(\mu_0 w_i, \lambda_0 w_i^2 )\,\! distribution for ''i'' = 1, 2, ..., ''n'' and all ''X''''i'' are independent, then : S=\sum_^n X_i \sim \operatorname\left( \mu_0 \sum w_i, \lambda_0 \left(\sum w_i \right)^2 \right). Note that : \frac= \frac =\frac is constant for all ''i''. This is a
necessary condition In logic and mathematics, necessity and sufficiency are terms used to describe a conditional or implicational relationship between two statements. For example, in the conditional statement: "If then ", is necessary for , because the truth of ...
for the summation. Otherwise ''S'' would not be Inverse Gaussian distributed.


Scaling

For any ''t'' > 0 it holds that : X \sim \operatorname(\mu,\lambda) \,\,\,\,\,\, \Rightarrow \,\,\,\,\,\, tX \sim \operatorname(t\mu,t\lambda).


Exponential family

The inverse Gaussian distribution is a two-parameter exponential family with
natural parameters In probability and statistics, an exponential family is a parametric set of probability distributions of a certain form, specified below. This special form is chosen for mathematical convenience, including the enabling of the user to calculat ...
−''λ''/(2''μ''2) and −''λ''/2, and
natural statistics In probability and statistics, an exponential family is a parametric set of probability distributions of a certain form, specified below. This special form is chosen for mathematical convenience, including the enabling of the user to calculate ...
''X'' and 1/''X''.


Relationship with Brownian motion

Let the
stochastic process In probability theory and related fields, a stochastic () or random process is a mathematical object usually defined as a family of random variables. Stochastic processes are widely used as mathematical models of systems and phenomena that appea ...
''X''''t'' be given by :X_0 = 0\quad :X_t = \nu t + \sigma W_t\quad\quad\quad\quad where ''W''''t'' is a standard Brownian motion. That is, ''X''''t'' is a Brownian motion with drift \nu > 0. Then the first passage time for a fixed level \alpha > 0 by ''X''''t'' is distributed according to an inverse-Gaussian: : T_\alpha = \inf\ \sim \operatorname \left(\frac\alpha\nu, \left(\frac \alpha \sigma \right)^2 \right) = \frac \exp\biggl(-\frac\biggr) i.e : P(T_ \in (T, T + dT)) = \frac \exp\biggl(-\frac\biggr)dT (cf. Schrödinger equation 19, Smoluchowski, equation 8, and Folks, equation 1). Suppose that we have a Brownian motion X_ with drift \nu defined by: :X_ = \nu t + \sigma W_, \quad X(0) = x_ And suppose that we wish to find the probability density function for the time when the process first hits some barrier \alpha > x_ - known as the first passage time. The Fokker-Planck equation describing the evolution of the probability distribution p(t,x) is: : + \nu = \sigma^, \quad \begin p(0,x) &= \delta(x-x_) \\ p(t,\alpha) &= 0 \end where \delta(\cdot) is the
Dirac delta function In mathematics, the Dirac delta distribution ( distribution), also known as the unit impulse, is a generalized function or distribution over the real numbers, whose value is zero everywhere except at zero, and whose integral over the entire ...
. This is a
boundary value problem In mathematics, in the field of differential equations, a boundary value problem is a differential equation together with a set of additional constraints, called the boundary conditions. A solution to a boundary value problem is a solution to t ...
(BVP) with a single absorbing boundary condition p(t,\alpha)=0, which may be solved using the method of images. Based on the initial condition, the fundamental solution to the Fokker-Planck equation, denoted by \varphi(t,x), is: : \varphi(t,x) = \exp\left - \right Define a point m, such that m>\alpha. This will allow the original and mirror solutions to cancel out exactly at the barrier at each instant in time. This implies that the initial condition should be augmented to become: : p(0,x) = \delta(x-x_) - A\delta(x-m) where A is a constant. Due to the linearity of the BVP, the solution to the Fokker-Planck equation with this initial condition is: : p(t,x) = \left\ Now we must determine the value of A. The fully absorbing boundary condition implies that: :(\alpha-x_-\nu t)^ = -2\sigma^t \log A + (\alpha - m - \nu t)^ At p(0,\alpha), we have that (\alpha-x_)^ = (\alpha-m)^ \implies m = 2\alpha - x_. Substituting this back into the above equation, we find that: :A = e^ Therefore, the full solution to the BVP is: :p(t,x) = \left\ Now that we have the full probability density function, we are ready to find the first passage time distribution f(t). The simplest route is to first compute the survival function S(t), which is defined as: :\begin S(t) &= \int_^p(t,x)dx \\ &= \Phi\left( \right ) - e^\Phi\left( \right ) \end where \Phi(\cdot) is the
cumulative distribution function In probability theory and statistics, the cumulative distribution function (CDF) of a real-valued random variable X, or just distribution function of X, evaluated at x, is the probability that X will take a value less than or equal to x. Ev ...
of the standard
normal distribution In statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real-valued random variable. The general form of its probability density function is : f(x) = \frac e^ The parameter \mu ...
. The survival function gives us the probability that the Brownian motion process has not crossed the barrier \alpha at some time t. Finally, the first passage time distribution f(t) is obtained from the identity: :\begin f(t) &= - \\ &= e^ \end Assuming that x_ = 0, the first passage time follows an inverse Gaussian distribution: :f(t) = e^ \sim \text\left ,\left( \right)^ \right/math>


When drift is zero

A common special case of the above arises when the Brownian motion has no drift. In that case, parameter ''μ'' tends to infinity, and the first passage time for fixed level ''α'' has probability density function : f \left( x; 0, \left(\frac \alpha \sigma \right)^2 \right) = \frac \alpha \exp\left(-\frac\right) (see also Bachelier). This is a Lévy distribution with parameters c=\left(\frac \alpha \sigma \right)^2 and \mu=0.


Maximum likelihood

The model where : X_i \sim \operatorname(\mu,\lambda w_i), \,\,\,\,\,\, i=1,2,\ldots,n with all ''w''''i'' known, (''μ'', ''λ'') unknown and all ''X''''i'' independent has the following likelihood function : L(\mu, \lambda)= \left( \frac \right)^\frac n 2 \left( \prod^n_ \frac \right)^ \exp\left(\frac \sum_^n w_i -\frac\sum_^n w_i X_i - \frac\lambda 2 \sum_^n w_i \frac1 \right). Solving the likelihood equation yields the following maximum likelihood estimates : \widehat= \frac, \,\,\,\,\,\,\,\, \frac= \frac \sum_^n w_i \left( \frac-\frac \right). \widehat and \widehat are independent and : \widehat \sim \operatorname \left(\mu, \lambda \sum_^n w_i \right), \qquad \frac \sim \frac \chi^2_.


Sampling from an inverse-Gaussian distribution

The following algorithm may be used.
Generate a random variate from a normal distribution with mean 0 and standard deviation equal 1 : \displaystyle \nu \sim N(0,1). Square the value : \displaystyle y = \nu^2 and use the relation : x = \mu + \frac - \frac\sqrt. Generate another random variate, this time sampled from a uniform distribution between 0 and 1 : \displaystyle z \sim U(0,1). If z \le \frac then return \displaystyle x else return \frac.
Sample code in Java: public double inverseGaussian(double mu, double lambda) And to plot Wald distribution in Python using
matplotlib Matplotlib is a plotting library for the Python programming language and its numerical mathematics extension NumPy. It provides an object-oriented API for embedding plots into applications using general-purpose GUI toolkits like Tkinter, wxPytho ...
and NumPy: import matplotlib.pyplot as plt import numpy as np h = plt.hist(np.random.wald(3, 2, 100000), bins=200, density=True) plt.show()


Related distributions

* If X \sim \operatorname(\mu,\lambda), then k X \sim \operatorname(k \mu,k \lambda) for any number k > 0. * If X_i \sim \operatorname(\mu,\lambda)\, then \sum_^n X_i \sim \operatorname(n \mu,n^2 \lambda)\, * If X_i \sim \operatorname(\mu,\lambda)\, for i=1,\ldots,n\, then \bar \sim \operatorname(\mu,n \lambda)\, * If X_i \sim \operatorname(\mu_i,2 \mu^2_i)\, then \sum_^n X_i \sim \operatorname\left(\sum_^n \mu_i, 2 \left( \sum_^n \mu_i \right)^2\right)\, * If X \sim \operatorname(\mu,\lambda), then \lambda (X-\mu)^2/\mu^2X \sim \chi^2(1). The convolution of an inverse Gaussian distribution (a Wald distribution) and an exponential (an ex-Wald distribution) is used as a model for response times in psychology, with visual search as one example.


History

This distribution appears to have been first derived in 1900 by Louis Bachelier as the time a stock reaches a certain price for the first time. In 1915 it was used independently by Erwin Schrödinger and Marian v. Smoluchowski as the time to first passage of a Brownian motion. In the field of reproduction modeling it is known as the Hadwiger function, after
Hugo Hadwiger Hugo Hadwiger (23 December 1908 in Karlsruhe, Germany – 29 October 1981 in Bern, Switzerland) was a Swiss mathematician, known for his work in geometry, combinatorics, and cryptography. Biography Although born in Karlsruhe, Germany, Hadwige ...
who described it in 1940. Abraham Wald re-derived this distribution in 1944 as the limiting form of a sample in a sequential probability ratio test. The name inverse Gaussian was proposed by
Maurice Tweedie Maurice Charles Kenneth Tweedie (born September 30, 1919 – died March 14, 1996), or Kenneth Tweedie, was a British medical physicist and statistician from the University of Liverpool. He was known for research into the exponential family proba ...
in 1945. Tweedie investigated this distribution in 1956 and 1957 and established some of its statistical properties. The distribution was extensively reviewed by Folks and Chhikara in 1978.


Numeric computation and software

Despite the simple formula for the probability density function, numerical probability calculations for the inverse Gaussian distribution nevertheless require special care to achieve full machine accuracy in floating point arithmetic for all parameter values. Functions for the inverse Gaussian distribution are provided for the R programming language by several packages including rmutil, SuppDists, STAR, invGauss, LaplacesDemon, and statmod.


See also

* Generalized inverse Gaussian distribution *
Tweedie distributions In probability and statistics, the Tweedie distributions are a family of probability distributions which include the purely continuous normal, gamma and inverse Gaussian distributions, the purely discrete scaled Poisson distribution, and the cla ...
—The inverse Gaussian distribution is a member of the family of Tweedie exponential dispersion models * Stopping time


References


Further reading

* *


External links


Inverse Gaussian Distribution
in Wolfram website. {{DEFAULTSORT:Inverse Gaussian Distribution Continuous distributions Exponential family distributions Infinitely divisible probability distributions Articles with example Java code Articles with example Python (programming language) code