HOME

TheInfoList



OR:

In
probability theory Probability theory is the branch of mathematics concerned with probability. Although there are several different probability interpretations, probability theory treats the concept in a rigorous mathematical manner by expressing it through a set o ...
, the inverse Gaussian distribution (also known as the Wald distribution) is a two-parameter family of
continuous probability distribution In probability theory and statistics, a probability distribution is the mathematical function that gives the probabilities of occurrence of different possible outcomes for an experiment. It is a mathematical description of a random phenomenon i ...
s with
support Support may refer to: Arts, entertainment, and media * Supporting character Business and finance * Support (technical analysis) * Child support * Customer support * Income Support Construction * Support (structure), or lateral support, a ...
on (0,∞). Its
probability density function In probability theory, a probability density function (PDF), or density of a continuous random variable, is a function whose value at any given sample (or point) in the sample space (the set of possible values taken by the random variable) can ...
is given by : f(x;\mu,\lambda) = \sqrt\frac \exp\biggl(-\frac\biggr) for ''x'' > 0, where \mu > 0 is the mean and \lambda > 0 is the shape parameter. The inverse Gaussian distribution has several properties analogous to a Gaussian distribution. The name can be misleading: it is an "inverse" only in that, while the Gaussian describes a Brownian motion's level at a fixed time, the inverse Gaussian describes the distribution of the time a Brownian motion with positive drift takes to reach a fixed positive level. Its cumulant generating function (logarithm of the characteristic function) is the inverse of the cumulant generating function of a Gaussian random variable. To indicate that a
random variable A random variable (also called random quantity, aleatory variable, or stochastic variable) is a mathematical formalization of a quantity or object which depends on random events. It is a mapping or a function from possible outcomes (e.g., the po ...
''X'' is inverse Gaussian-distributed with mean μ and shape parameter λ we write X \sim \operatorname(\mu, \lambda)\,\!.


Properties


Single parameter form

The probability density function (pdf) of the inverse Gaussian distribution has a single parameter form given by : f(x;\mu,\mu^2) = \frac \exp\biggl(-\frac\biggr). In this form, the mean and variance of the distribution are equal, \mathbb = \text(X). Also, the cumulative distribution function (cdf) of the single parameter inverse Gaussian distribution is related to the standard normal distribution by : \begin \Pr(X < x) &= \Phi(z_1) + e^ \Phi(z_2), & \text & \quad 0 < x \leq \mu, \\ \Pr(X > x) &= \Phi(-z_1) - e^ \Phi(z_2), & \text & \quad x \geq \mu. \end where z_1 = \frac - x^, z_2 = \frac + x^, and the \Phi is the cdf of standard normal distribution. The variables z_1 and z_2 are related to each other by the identity z_2^2 = z_1^2 + 4\mu. In the single parameter form, the MGF simplifies to : M(t) = \exp mu(1-\sqrt) An inverse Gaussian distribution in double parameter form f(x;\mu,\lambda) can be transformed into a single parameter form f(y;\mu_0,\mu_0^2) by appropriate scaling y = \frac, where \mu_0 = \mu^3/\lambda. The standard form of inverse Gaussian distribution is : f(x;1,1) = \frac \exp\biggl(-\frac\biggr).


Summation

If ''X''''i'' has an \operatorname(\mu_0 w_i, \lambda_0 w_i^2 )\,\! distribution for ''i'' = 1, 2, ..., ''n'' and all ''X''''i'' are
independent Independent or Independents may refer to: Arts, entertainment, and media Artist groups * Independents (artist group), a group of modernist painters based in the New Hope, Pennsylvania, area of the United States during the early 1930s * Independ ...
, then : S=\sum_^n X_i \sim \operatorname\left( \mu_0 \sum w_i, \lambda_0 \left(\sum w_i \right)^2 \right). Note that : \frac= \frac =\frac is constant for all ''i''. This is a
necessary condition In logic and mathematics, necessity and sufficiency are terms used to describe a conditional or implicational relationship between two statements. For example, in the conditional statement: "If then ", is necessary for , because the truth of ...
for the summation. Otherwise ''S'' would not be Inverse Gaussian distributed.


Scaling

For any ''t'' > 0 it holds that : X \sim \operatorname(\mu,\lambda) \,\,\,\,\,\, \Rightarrow \,\,\,\,\,\, tX \sim \operatorname(t\mu,t\lambda).


Exponential family

The inverse Gaussian distribution is a two-parameter
exponential family In probability and statistics, an exponential family is a parametric set of probability distributions of a certain form, specified below. This special form is chosen for mathematical convenience, including the enabling of the user to calculate ...
with
natural parameters In probability and statistics, an exponential family is a parametric set of probability distributions of a certain form, specified below. This special form is chosen for mathematical convenience, including the enabling of the user to calculat ...
−''λ''/(2''μ''2) and −''λ''/2, and
natural statistics In theory of probability, probability and statistics, an exponential family is a parametric model, parametric set of probability distributions of a certain form, specified below. This special form is chosen for mathematical convenience, includin ...
''X'' and 1/''X''.


Relationship with Brownian motion

Let the
stochastic process In probability theory and related fields, a stochastic () or random process is a mathematical object usually defined as a family of random variables. Stochastic processes are widely used as mathematical models of systems and phenomena that appea ...
''X''''t'' be given by :X_0 = 0\quad :X_t = \nu t + \sigma W_t\quad\quad\quad\quad where ''W''''t'' is a standard
Brownian motion Brownian motion, or pedesis (from grc, πήδησις "leaping"), is the random motion of particles suspended in a medium (a liquid or a gas). This pattern of motion typically consists of random fluctuations in a particle's position insi ...
. That is, ''X''''t'' is a Brownian motion with drift \nu > 0. Then the
first passage time Events are often triggered when a stochastic or random process first encounters a threshold. The threshold can be a barrier, boundary or specified state of a system. The amount of time required for a stochastic process, starting from some initial ...
for a fixed level \alpha > 0 by ''X''''t'' is distributed according to an inverse-Gaussian: : T_\alpha = \inf\ \sim \operatorname \left(\frac\alpha\nu, \left(\frac \alpha \sigma \right)^2 \right) = \frac \exp\biggl(-\frac\biggr) i.e : P(T_ \in (T, T + dT)) = \frac \exp\biggl(-\frac\biggr)dT (cf. Schrödinger equation 19, Smoluchowski, equation 8, and Folks, equation 1). Suppose that we have a Brownian motion X_ with drift \nu defined by: :X_ = \nu t + \sigma W_, \quad X(0) = x_ And suppose that we wish to find the
probability density function In probability theory, a probability density function (PDF), or density of a continuous random variable, is a function whose value at any given sample (or point) in the sample space (the set of possible values taken by the random variable) can ...
for the time when the process first hits some barrier \alpha > x_ - known as the first passage time. The Fokker-Planck equation describing the evolution of the probability distribution p(t,x) is: : + \nu = \sigma^, \quad \begin p(0,x) &= \delta(x-x_) \\ p(t,\alpha) &= 0 \end where \delta(\cdot) is the
Dirac delta function In mathematics, the Dirac delta distribution ( distribution), also known as the unit impulse, is a generalized function or distribution over the real numbers, whose value is zero everywhere except at zero, and whose integral over the entire ...
. This is a
boundary value problem In mathematics, in the field of differential equations, a boundary value problem is a differential equation together with a set of additional constraints, called the boundary conditions. A solution to a boundary value problem is a solution to t ...
(BVP) with a single absorbing boundary condition p(t,\alpha)=0, which may be solved using the
method of images The method of images (or method of mirror images) is a mathematical tool for solving differential equations, in which the domain of the sought function is extended by the addition of its mirror image with respect to a symmetry hyperplane. As a resul ...
. Based on the initial condition, the
fundamental solution In mathematics, a fundamental solution for a linear partial differential operator is a formulation in the language of distribution theory of the older idea of a Green's function (although unlike Green's functions, fundamental solutions do not ad ...
to the Fokker-Planck equation, denoted by \varphi(t,x), is: : \varphi(t,x) = \exp\left - \right Define a point m, such that m>\alpha. This will allow the original and mirror solutions to cancel out exactly at the barrier at each instant in time. This implies that the initial condition should be augmented to become: : p(0,x) = \delta(x-x_) - A\delta(x-m) where A is a constant. Due to the linearity of the BVP, the solution to the Fokker-Planck equation with this initial condition is: : p(t,x) = \left\ Now we must determine the value of A. The fully absorbing boundary condition implies that: :(\alpha-x_-\nu t)^ = -2\sigma^t \log A + (\alpha - m - \nu t)^ At p(0,\alpha), we have that (\alpha-x_)^ = (\alpha-m)^ \implies m = 2\alpha - x_. Substituting this back into the above equation, we find that: :A = e^ Therefore, the full solution to the BVP is: :p(t,x) = \left\ Now that we have the full probability density function, we are ready to find the first passage time distribution f(t). The simplest route is to first compute the
survival function The survival function is a function that gives the probability that a patient, device, or other object of interest will survive past a certain time. The survival function is also known as the survivor function or reliability function. The term ...
S(t), which is defined as: :\begin S(t) &= \int_^p(t,x)dx \\ &= \Phi\left( \right ) - e^\Phi\left( \right ) \end where \Phi(\cdot) is the
cumulative distribution function In probability theory and statistics, the cumulative distribution function (CDF) of a real-valued random variable X, or just distribution function of X, evaluated at x, is the probability that X will take a value less than or equal to x. Ev ...
of the standard
normal distribution In statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real-valued random variable. The general form of its probability density function is : f(x) = \frac e^ The parameter \mu ...
. The survival function gives us the probability that the Brownian motion process has not crossed the barrier \alpha at some time t. Finally, the first passage time distribution f(t) is obtained from the identity: :\begin f(t) &= - \\ &= e^ \end Assuming that x_ = 0, the first passage time follows an inverse Gaussian distribution: :f(t) = e^ \sim \text\left ,\left( \right)^ \right/math>


When drift is zero

A common special case of the above arises when the Brownian motion has no drift. In that case, parameter ''μ'' tends to infinity, and the first passage time for fixed level ''α'' has probability density function : f \left( x; 0, \left(\frac \alpha \sigma \right)^2 \right) = \frac \alpha \exp\left(-\frac\right) (see also Bachelier). This is a
Lévy distribution In probability theory and statistics, the Lévy distribution, named after Paul Lévy, is a continuous probability distribution for a non-negative random variable. In spectroscopy, this distribution, with frequency as the dependent variable, is k ...
with parameters c=\left(\frac \alpha \sigma \right)^2 and \mu=0.


Maximum likelihood

The model where : X_i \sim \operatorname(\mu,\lambda w_i), \,\,\,\,\,\, i=1,2,\ldots,n with all ''w''''i'' known, (''μ'', ''λ'') unknown and all ''X''''i''
independent Independent or Independents may refer to: Arts, entertainment, and media Artist groups * Independents (artist group), a group of modernist painters based in the New Hope, Pennsylvania, area of the United States during the early 1930s * Independ ...
has the following likelihood function : L(\mu, \lambda)= \left( \frac \right)^\frac n 2 \left( \prod^n_ \frac \right)^ \exp\left(\frac \sum_^n w_i -\frac\sum_^n w_i X_i - \frac\lambda 2 \sum_^n w_i \frac1 \right). Solving the likelihood equation yields the following maximum likelihood estimates : \widehat= \frac, \,\,\,\,\,\,\,\, \frac= \frac \sum_^n w_i \left( \frac-\frac \right). \widehat and \widehat are independent and : \widehat \sim \operatorname \left(\mu, \lambda \sum_^n w_i \right), \qquad \frac \sim \frac \chi^2_.


Sampling from an inverse-Gaussian distribution

The following algorithm may be used.
Generate a random variate from a normal distribution with mean 0 and standard deviation equal 1 : \displaystyle \nu \sim N(0,1). Square the value : \displaystyle y = \nu^2 and use the relation : x = \mu + \frac - \frac\sqrt. Generate another random variate, this time sampled from a uniform distribution between 0 and 1 : \displaystyle z \sim U(0,1). If z \le \frac then return \displaystyle x else return \frac.
Sample code in
Java Java (; id, Jawa, ; jv, ꦗꦮ; su, ) is one of the Greater Sunda Islands in Indonesia. It is bordered by the Indian Ocean to the south and the Java Sea to the north. With a population of 151.6 million people, Java is the world's List ...
: public double inverseGaussian(double mu, double lambda) And to plot Wald distribution in
Python Python may refer to: Snakes * Pythonidae, a family of nonvenomous snakes found in Africa, Asia, and Australia ** ''Python'' (genus), a genus of Pythonidae found in Africa and Asia * Python (mythology), a mythical serpent Computing * Python (pro ...
using matplotlib and NumPy: import matplotlib.pyplot as plt import numpy as np h = plt.hist(np.random.wald(3, 2, 100000), bins=200, density=True) plt.show()


Related distributions

* If X \sim \operatorname(\mu,\lambda), then k X \sim \operatorname(k \mu,k \lambda) for any number k > 0. * If X_i \sim \operatorname(\mu,\lambda)\, then \sum_^n X_i \sim \operatorname(n \mu,n^2 \lambda)\, * If X_i \sim \operatorname(\mu,\lambda)\, for i=1,\ldots,n\, then \bar \sim \operatorname(\mu,n \lambda)\, * If X_i \sim \operatorname(\mu_i,2 \mu^2_i)\, then \sum_^n X_i \sim \operatorname\left(\sum_^n \mu_i, 2 \left( \sum_^n \mu_i \right)^2\right)\, * If X \sim \operatorname(\mu,\lambda), then \lambda (X-\mu)^2/\mu^2X \sim \chi^2(1). The convolution of an inverse Gaussian distribution (a Wald distribution) and an exponential (an ex-Wald distribution) is used as a model for response times in psychology, with visual search as one example.


History

This distribution appears to have been first derived in 1900 by
Louis Bachelier Louis Jean-Baptiste Alphonse Bachelier (; 11 March 1870 – 28 April 1946) was a French mathematician at the turn of the 20th century. He is credited with being the first person to model the stochastic process now called Brownian motion, as part ...
as the time a stock reaches a certain price for the first time. In 1915 it was used independently by
Erwin Schrödinger Erwin Rudolf Josef Alexander Schrödinger (, ; ; 12 August 1887 – 4 January 1961), sometimes written as or , was a Nobel Prize-winning Austrian physicist with Irish citizenship who developed a number of fundamental results in quantum theory ...
and Marian v. Smoluchowski as the time to first passage of a Brownian motion. In the field of reproduction modeling it is known as the Hadwiger function, after
Hugo Hadwiger Hugo Hadwiger (23 December 1908 in Karlsruhe, Germany – 29 October 1981 in Bern, Switzerland) was a Swiss mathematician, known for his work in geometry, combinatorics, and cryptography. Biography Although born in Karlsruhe, Germany, Hadwi ...
who described it in 1940.
Abraham Wald Abraham Wald (; hu, Wald Ábrahám, yi, אברהם וואַלד;  – ) was a Jewish Hungarian mathematician who contributed to decision theory, geometry, and econometrics and founded the field of statistical sequential analysis. One of ...
re-derived this distribution in 1944 as the limiting form of a sample in a sequential probability ratio test. The name inverse Gaussian was proposed by Maurice Tweedie in 1945. Tweedie investigated this distribution in 1956 and 1957 and established some of its statistical properties. The distribution was extensively reviewed by Folks and Chhikara in 1978.


Numeric computation and software

Despite the simple formula for the probability density function, numerical probability calculations for the inverse Gaussian distribution nevertheless require special care to achieve full machine accuracy in floating point arithmetic for all parameter values. Functions for the inverse Gaussian distribution are provided for the
R programming language R is a programming language for statistical computing and graphics supported by the R Core Team and the R Foundation for Statistical Computing. Created by statisticians Ross Ihaka and Robert Gentleman, R is used among data miners, bioinform ...
by several packages including rmutil, SuppDists, STAR, invGauss, LaplacesDemon, and statmod.


See also

*
Generalized inverse Gaussian distribution In probability theory and statistics, the generalized inverse Gaussian distribution (GIG) is a three-parameter family of continuous probability distributions with probability density function :f(x) = \frac x^ e^,\qquad x>0, where ''Kp'' is a mo ...
*
Tweedie distributions In probability and statistics, the Tweedie distributions are a family of probability distributions which include the purely continuous normal, gamma and inverse Gaussian distributions, the purely discrete scaled Poisson distribution, and the cla ...
—The inverse Gaussian distribution is a member of the family of Tweedie
exponential dispersion model In probability and statistics, the class of exponential dispersion models (EDM) is a set of probability distributions that represents a generalisation of the natural exponential family.Jørgensen, B. (1987). Exponential dispersion models (with dis ...
s *
Stopping time In probability theory, in particular in the study of stochastic processes, a stopping time (also Markov time, Markov moment, optional stopping time or optional time ) is a specific type of “random time”: a random variable whose value is inter ...


References


Further reading

* *


External links


Inverse Gaussian Distribution
in Wolfram website. {{DEFAULTSORT:Inverse Gaussian Distribution Continuous distributions Exponential family distributions Infinitely divisible probability distributions Articles with example Java code Articles with example Python (programming language) code