HOME

TheInfoList



OR:

In
probability theory Probability theory is the branch of mathematics concerned with probability. Although there are several different probability interpretations, probability theory treats the concept in a rigorous mathematical manner by expressing it through a set ...
and statistics, the normal-inverse-gamma distribution (or Gaussian-inverse-gamma distribution) is a four-parameter family of multivariate continuous probability distributions. It is the
conjugate prior In Bayesian probability theory, if the posterior distribution p(\theta \mid x) is in the same probability distribution family as the prior probability distribution p(\theta), the prior and posterior are then called conjugate distributions, and th ...
of a
normal distribution In statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real-valued random variable. The general form of its probability density function is : f(x) = \frac e^ The parameter \mu ...
with unknown
mean There are several kinds of mean in mathematics, especially in statistics. Each mean serves to summarize a given group of data, often to better understand the overall value (magnitude and sign) of a given data set. For a data set, the '' ari ...
and
variance In probability theory and statistics, variance is the expectation of the squared deviation of a random variable from its population mean or sample mean. Variance is a measure of dispersion, meaning it is a measure of how far a set of numbe ...
.


Definition

Suppose : x \mid \sigma^2, \mu, \lambda\sim \mathrm(\mu,\sigma^2 / \lambda) \,\! has a
normal distribution In statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real-valued random variable. The general form of its probability density function is : f(x) = \frac e^ The parameter \mu ...
with
mean There are several kinds of mean in mathematics, especially in statistics. Each mean serves to summarize a given group of data, often to better understand the overall value (magnitude and sign) of a given data set. For a data set, the '' ari ...
\mu and
variance In probability theory and statistics, variance is the expectation of the squared deviation of a random variable from its population mean or sample mean. Variance is a measure of dispersion, meaning it is a measure of how far a set of numbe ...
\sigma^2 / \lambda, where :\sigma^2\mid\alpha, \beta \sim \Gamma^(\alpha,\beta) \! has an inverse gamma distribution. Then (x,\sigma^2) has a normal-inverse-gamma distribution, denoted as : (x,\sigma^2) \sim \text\Gamma^(\mu,\lambda,\alpha,\beta) \! . (\text is also used instead of \text\Gamma^.) The
normal-inverse-Wishart distribution In probability theory and statistics, the normal-inverse-Wishart distribution (or Gaussian-inverse-Wishart distribution) is a multivariate four-parameter family of continuous probability distributions. It is the conjugate prior of a multivariate n ...
is a generalization of the normal-inverse-gamma distribution that is defined over multivariate random variables.


Characterization


Probability density function

: f(x,\sigma^2\mid\mu,\lambda,\alpha,\beta) = \frac \, \frac \, \left( \frac \right)^ \exp \left( -\frac \right) For the multivariate form where \mathbf is a k \times 1 random vector, : f(\mathbf,\sigma^2\mid\mu,\mathbf^,\alpha,\beta) = , \mathbf, ^ \, \frac \, \left( \frac \right)^ \exp \left( -\frac \right). where , \mathbf, is the
determinant In mathematics, the determinant is a scalar value that is a function of the entries of a square matrix. It characterizes some properties of the matrix and the linear map represented by the matrix. In particular, the determinant is nonzero if a ...
of the k \times k
matrix Matrix most commonly refers to: * ''The Matrix'' (franchise), an American media franchise ** ''The Matrix'', a 1999 science-fiction action film ** "The Matrix", a fictional setting, a virtual reality environment, within ''The Matrix'' (franchis ...
\mathbf. Note how this last equation reduces to the first form if k = 1 so that \mathbf, \mathbf, \boldsymbol are scalars.


Alternative parameterization

It is also possible to let \gamma = 1 / \lambda in which case the pdf becomes : f(x,\sigma^2\mid\mu,\gamma,\alpha,\beta) = \frac \, \frac \, \left( \frac \right)^ \exp \left( -\frac \right) In the multivariate form, the corresponding change would be to regard the covariance matrix \mathbf instead of its inverse \mathbf^ as a parameter.


Cumulative distribution function

: F(x,\sigma^2\mid\mu,\lambda,\alpha,\beta) = \frac


Properties


Marginal distributions

Given (x,\sigma^2) \sim \text\Gamma^(\mu,\lambda,\alpha,\beta) \! . as above, \sigma^2 by itself follows an inverse gamma distribution: :\sigma^2 \sim \Gamma^(\alpha,\beta) \! while \sqrt (x - \mu) follows a
t distribution The phrase "T distribution" may refer to * Student's t-distribution in univariate probability theory, * Hotelling's T-square distribution in multivariate statistics. * Multivariate Student distribution In statistics, the multivariate ''t''-dis ...
with 2 \alpha degrees of freedom. In the multivariate case, the marginal distribution of \mathbf is a multivariate t distribution: :\mathbf \sim t_(\boldsymbol, \frac \mathbf^) \!


Summation


Scaling

Suppose : (x,\sigma^2) \sim \text\Gamma^(\mu,\lambda,\alpha,\beta) \! . Then for c>0 , : (cx,c\sigma^2) \sim \text\Gamma^(c\mu,\lambda/c,\alpha,c\beta) \! . Proof: To prove this let (x,\sigma^2) \sim \text\Gamma^(\mu,\lambda,\alpha,\beta) and fix c>0 . Defining Y=(Y_1,Y_2)=(cx,c \sigma^2) , observe that the PDF of the random variable Y evaluated at (y_1,y_2) is given by 1/c^2 times the PDF of a \text\Gamma^(\mu,\lambda,\alpha,\beta) random variable evaluated at (y_1/c,y_2/c) . Hence the PDF of Y evaluated at (y_1,y_2) is given by : f_Y(y_1,y_2)=\frac \frac \, \frac \, \left( \frac \right)^ \exp \left( -\frac \right) = \frac \, \frac \, \left( \frac \right)^ \exp \left( -\frac \right).\! The right hand expression is the PDF for a \text\Gamma^(c\mu,\lambda/c,\alpha,c\beta) random variable evaluated at (y_1,y_2) , which completes the proof.


Exponential family

Normal distributions form an
exponential family In probability and statistics, an exponential family is a parametric set of probability distributions of a certain form, specified below. This special form is chosen for mathematical convenience, including the enabling of the user to calculate ...
with
natural parameter In probability and statistics, an exponential family is a parametric set of probability distributions of a certain form, specified below. This special form is chosen for mathematical convenience, including the enabling of the user to calculate ...
s \textstyle\theta_1=\frac, \textstyle\theta_2=\lambda \mu, \textstyle\theta_3=\alpha , and \textstyle\theta_4=-\beta+\frac and sufficient statistics \textstyle T_1=\frac, \textstyle T_2=\frac, \textstyle T_3=\log \big( \frac \big) , and \textstyle T_4=\frac.


Information entropy


Kullback–Leibler divergence

Measures difference between two distributions.


Maximum likelihood estimation


Posterior distribution of the parameters

See the articles on
normal-gamma distribution In probability theory and statistics, the normal-gamma distribution (or Gaussian-gamma distribution) is a bivariate four-parameter family of continuous probability distributions. It is the conjugate prior of a normal distribution with unknown mean ...
and
conjugate prior In Bayesian probability theory, if the posterior distribution p(\theta \mid x) is in the same probability distribution family as the prior probability distribution p(\theta), the prior and posterior are then called conjugate distributions, and th ...
.


Interpretation of the parameters

See the articles on
normal-gamma distribution In probability theory and statistics, the normal-gamma distribution (or Gaussian-gamma distribution) is a bivariate four-parameter family of continuous probability distributions. It is the conjugate prior of a normal distribution with unknown mean ...
and
conjugate prior In Bayesian probability theory, if the posterior distribution p(\theta \mid x) is in the same probability distribution family as the prior probability distribution p(\theta), the prior and posterior are then called conjugate distributions, and th ...
.


Generating normal-inverse-gamma random variates

Generation of random variates is straightforward: # Sample \sigma^2 from an inverse gamma distribution with parameters \alpha and \beta # Sample x from a normal distribution with mean \mu and variance \sigma^2/\lambda


Related distributions

* The
normal-gamma distribution In probability theory and statistics, the normal-gamma distribution (or Gaussian-gamma distribution) is a bivariate four-parameter family of continuous probability distributions. It is the conjugate prior of a normal distribution with unknown mean ...
is the same distribution parameterized by
precision Precision, precise or precisely may refer to: Science, and technology, and mathematics Mathematics and computing (general) * Accuracy and precision, measurement deviation from true value and its scatter * Significant figures, the number of digit ...
rather than
variance In probability theory and statistics, variance is the expectation of the squared deviation of a random variable from its population mean or sample mean. Variance is a measure of dispersion, meaning it is a measure of how far a set of numbe ...
* A generalization of this distribution which allows for a multivariate mean and a completely unknown positive-definite covariance matrix \sigma^2 \mathbf (whereas in the multivariate inverse-gamma distribution the covariance matrix is regarded as known up to the scale factor \sigma^2) is the
normal-inverse-Wishart distribution In probability theory and statistics, the normal-inverse-Wishart distribution (or Gaussian-inverse-Wishart distribution) is a multivariate four-parameter family of continuous probability distributions. It is the conjugate prior of a multivariate n ...


See also

*
Compound probability distribution In probability and statistics, a compound probability distribution (also known as a mixture distribution or contagious distribution) is the probability distribution that results from assuming that a random variable is distributed according to som ...


References

* Denison, David G. T. ; Holmes, Christopher C.; Mallick, Bani K.; Smith, Adrian F. M. (2002) ''Bayesian Methods for Nonlinear Classification and Regression'', Wiley. * Koch, Karl-Rudolf (2007) ''Introduction to Bayesian Statistics'' (2nd Edition), Springer. {{ProbDistributions, multivariate Continuous distributions Multivariate continuous distributions Normal distribution