HOME

TheInfoList



OR:

In statistics, the Wishart distribution is a generalization to multiple dimensions of the
gamma distribution In probability theory and statistics, the gamma distribution is a two- parameter family of continuous probability distributions. The exponential distribution, Erlang distribution, and chi-square distribution are special cases of the gamma dis ...
. It is named in honor of John Wishart, who first formulated the distribution in 1928. It is a family of
probability distribution In probability theory and statistics, a probability distribution is the mathematical function that gives the probabilities of occurrence of different possible outcomes for an experiment. It is a mathematical description of a random phenomeno ...
s defined over symmetric,
nonnegative-definite In mathematics, a symmetric matrix M with real entries is positive-definite if the real number z^\textsfMz is positive for every nonzero real column vector z, where z^\textsf is the transpose of More generally, a Hermitian matrix (that is, ...
random matrices (i.e.
matrix Matrix most commonly refers to: * ''The Matrix'' (franchise), an American media franchise ** '' The Matrix'', a 1999 science-fiction action film ** "The Matrix", a fictional setting, a virtual reality environment, within ''The Matrix'' (franchi ...
-valued
random variable A random variable (also called random quantity, aleatory variable, or stochastic variable) is a mathematical formalization of a quantity or object which depends on random events. It is a mapping or a function from possible outcomes (e.g., the p ...
s). In random matrix theory, the space of Wishart matrices is called the ''Wishart ensemble''. These distributions are of great importance in the estimation of covariance matrices in
multivariate statistics Multivariate statistics is a subdivision of statistics encompassing the simultaneous observation and analysis of more than one outcome variable. Multivariate statistics concerns understanding the different aims and background of each of the dif ...
. In
Bayesian statistics Bayesian statistics is a theory in the field of statistics based on the Bayesian interpretation of probability where probability expresses a ''degree of belief'' in an event. The degree of belief may be based on prior knowledge about the event, ...
, the Wishart distribution is the
conjugate prior In Bayesian probability theory, if the posterior distribution p(\theta \mid x) is in the same probability distribution family as the prior probability distribution p(\theta), the prior and posterior are then called conjugate distributions, and t ...
of the
inverse Inverse or invert may refer to: Science and mathematics * Inverse (logic), a type of conditional sentence which is an immediate inference made from another conditional sentence * Additive inverse (negation), the inverse of a number that, when ad ...
covariance-matrix of a multivariate-normal random-vector.


Definition

Suppose is a matrix, each column of which is independently drawn from a -variate normal distribution with zero mean: :G_ = (g_i^1,\dots,g_i^p)^T\sim \mathcal_p(0,V). Then the Wishart distribution is the
probability distribution In probability theory and statistics, a probability distribution is the mathematical function that gives the probabilities of occurrence of different possible outcomes for an experiment. It is a mathematical description of a random phenomeno ...
of the random matrix :S= G G^T = \sum_^n G_G_^T known as the scatter matrix. One indicates that has that probability distribution by writing :S\sim W_p(V,n). The positive integer is the number of ''
degrees of freedom Degrees of freedom (often abbreviated df or DOF) refers to the number of independent variables or parameters of a thermodynamic system. In various scientific fields, the word "freedom" is used to describe the limits to which physical movement or ...
''. Sometimes this is written . For the matrix is invertible with probability if is invertible. If then this distribution is a
chi-squared distribution In probability theory and statistics, the chi-squared distribution (also chi-square or \chi^2-distribution) with k degrees of freedom is the distribution of a sum of the squares of k independent standard normal random variables. The chi-squar ...
with degrees of freedom.


Occurrence

The Wishart distribution arises as the distribution of the sample covariance matrix for a sample from a
multivariate normal distribution In probability theory and statistics, the multivariate normal distribution, multivariate Gaussian distribution, or joint normal distribution is a generalization of the one-dimensional ( univariate) normal distribution to higher dimensions. One ...
. It occurs frequently in
likelihood-ratio test In statistics, the likelihood-ratio test assesses the goodness of fit of two competing statistical models based on the ratio of their likelihoods, specifically one found by maximization over the entire parameter space and another found after ...
s in multivariate statistical analysis. It also arises in the spectral theory of random matrices and in multidimensional Bayesian analysis. It is also encountered in wireless communications, while analyzing the performance of
Rayleigh fading Rayleigh fading is a statistical model for the effect of a propagation environment on a radio signal, such as that used by wireless devices. Rayleigh fading models assume that the magnitude of a signal that has passed through such a transmissio ...
MIMO In radio, multiple-input and multiple-output, or MIMO (), is a method for multiplying the capacity of a radio link using multiple transmission and receiving antennas to exploit multipath propagation. MIMO has become an essential element of w ...
wireless channels .


Probability density function

The Wishart distribution can be characterized by its
probability density function In probability theory, a probability density function (PDF), or density of a continuous random variable, is a function whose value at any given sample (or point) in the sample space (the set of possible values taken by the random variable) c ...
as follows: Let be a symmetric matrix of random variables that is positive semi-definite. Let be a (fixed) symmetric positive definite matrix of size . Then, if , has a Wishart distribution with degrees of freedom if it has the
probability density function In probability theory, a probability density function (PDF), or density of a continuous random variable, is a function whose value at any given sample (or point) in the sample space (the set of possible values taken by the random variable) c ...
: f_ (\mathbf x) = \frac^ e^ where \left, \ is the
determinant In mathematics, the determinant is a scalar value that is a function of the entries of a square matrix. It characterizes some properties of the matrix and the linear map represented by the matrix. In particular, the determinant is nonzero if ...
of \mathbf x and is the
multivariate gamma function In mathematics, the multivariate gamma function Γ''p'' is a generalization of the gamma function. It is useful in multivariate statistics, appearing in the probability density function of the Wishart and inverse Wishart distributions, and the mat ...
defined as :\Gamma_p \left (\frac n 2 \right )= \pi^\prod_^p \Gamma\left( \frac - \frac \right ). The density above is not the joint density of all the p^2 elements of the random matrix (such density does not exist because of the symmetry constrains X_=X_), it is rather the joint density of p(p+1)/2 elements X_ for i\le j (, page 38). Also, the density formula above applies only to positive definite matrices \mathbf x; for other matrices the density is equal to zero. The joint-eigenvalue density for the eigenvalues \lambda_1,\dots , \lambda_p\ge 0 of a random matrix \mathbf\sim W_p(\mathbf,n) is, : c_e^\prod \lambda_i^\prod_, \lambda_i-\lambda_j, where c_is a constant. In fact the above definition can be extended to any real . If , then the Wishart no longer has a density—instead it represents a singular distribution that takes values in a lower-dimension subspace of the space of matrices.


Use in Bayesian statistics

In
Bayesian statistics Bayesian statistics is a theory in the field of statistics based on the Bayesian interpretation of probability where probability expresses a ''degree of belief'' in an event. The degree of belief may be based on prior knowledge about the event, ...
, in the context of the
multivariate normal distribution In probability theory and statistics, the multivariate normal distribution, multivariate Gaussian distribution, or joint normal distribution is a generalization of the one-dimensional ( univariate) normal distribution to higher dimensions. One ...
, the Wishart distribution is the conjugate prior to the precision matrix , where is the covariance matrix.


Choice of parameters

The least informative, proper Wishart prior is obtained by setting . The prior mean of is , suggesting that a reasonable choice for would be , where is some prior guess for the covariance matrix.


Properties


Log-expectation

The following formula plays a role in variational Bayes derivations for Bayes networks involving the Wishart distribution: :\operatorname \mathbf\\, = \psi_p\left(\frac n 2\right) + p \, \ln(2) + \ln, \mathbf, where \psi_p is the multivariate digamma function (the derivative of the log of the
multivariate gamma function In mathematics, the multivariate gamma function Γ''p'' is a generalization of the gamma function. It is useful in multivariate statistics, appearing in the probability density function of the Wishart and inverse Wishart distributions, and the mat ...
).


Log-variance

The following variance computation could be of help in Bayesian statistics: : \operatorname\left \mathbf\ \,\right\sum_^p \psi_1\left(\frac 2\right) where \psi_1 is the trigamma function. This comes up when computing the Fisher information of the Wishart random variable.


Entropy

The
information entropy In information theory, the entropy of a random variable is the average level of "information", "surprise", or "uncertainty" inherent to the variable's possible outcomes. Given a discrete random variable X, which takes values in the alphabet \ ...
of the distribution has the following formula: :\operatorname\left , \mathbf \,\right= -\ln \left( B(\mathbf,n) \right) -\frac \operatorname\left \mathbf\\,\right+ \frac where is the normalizing constant of the distribution: :B(\mathbf,n) = \frac. This can be expanded as follows: : \begin \operatorname\left , \mathbf\, \right& = \frac \ln \left, \mathbf\ +\frac \ln 2 + \ln \Gamma_p \left(\frac \right) - \frac \operatorname\left \mathbf\\, \right+ \frac \\ pt&= \frac \ln\left, \mathbf\ + \frac \ln 2 + \ln\Gamma_p\left(\frac \right) - \frac 2 \left( \psi_p \left(\frac\right) + p\ln 2 + \ln\left, \mathbf\\right) + \frac \\ pt&= \frac \ln\left, \mathbf\ + \frac \ln 2 + \ln\Gamma_p\left(\frac n 2\right) - \frac \psi_p\left(\frac n 2 \right) - \frac 2 \left(p\ln 2 +\ln\left, \mathbf\ \right) + \frac \\ pt&= \frac \ln\left, \mathbf\ + \frac1 2 p(p+1) \ln 2 + \ln\Gamma_p\left(\frac n 2\right) - \frac \psi_p\left(\frac n 2 \right) + \frac \end


Cross-entropy

The cross entropy of two Wishart distributions p_0 with parameters n_0, V_0 and p_1 with parameters n_1, V_1 is :\begin H(p_0, p_1) &= \operatorname_ , -\log p_1\, \ pt&= \operatorname_ \left , -\log \frac \right\ pt&= \tfrac 2 \log 2 + \tfrac 2 \log \left, \mathbf_1\ + \log \Gamma_(\tfrac 2) - \tfrac 2 \operatorname_\left \mathbf\\, \right+ \tfrac\operatorname_\left , \operatorname\left(\,\mathbf_1^\mathbf\,\right) \, \right\\ pt&= \tfrac \log 2 + \tfrac 2 \log \left, \mathbf_1\ + \log \Gamma_(\tfrac) - \tfrac \left( \psi_(\tfrac 2) + p_0 \log 2 + \log \left, \mathbf_0\\right)+ \tfrac \operatorname\left(\, \mathbf_1^ n_0 \mathbf_0\, \right) \\ pt&=-\tfrac \log \left, \, \mathbf_1^ \mathbf_0\, \ + \tfrac 2 \log \left, \mathbf_0\ + \tfrac 2 \operatorname\left(\, \mathbf_1^ \mathbf_0\right)+ \log \Gamma_\left(\tfrac\right) - \tfrac \psi_(\tfrac) + \tfrac \log 2 \end Note that when p_0=p_1 and n_0=n_1we recover the entropy.


KL-divergence

The
Kullback–Leibler divergence In mathematical statistics, the Kullback–Leibler divergence (also called relative entropy and I-divergence), denoted D_\text(P \parallel Q), is a type of statistical distance: a measure of how one probability distribution ''P'' is different fro ...
of p_1 from p_0 is : \begin D_(p_0 \, p_1) & = H(p_0, p_1) - H(p_0) \\ pt & =-\frac 2 \log , \mathbf_1^ \mathbf_0, + \frac(\operatorname(\mathbf_1^ \mathbf_0) - p)+ \log \frac + \tfrac 2 \psi_p\left(\frac 2\right) \end


Characteristic function

The
characteristic function In mathematics, the term "characteristic function" can refer to any of several distinct concepts: * The indicator function of a subset, that is the function ::\mathbf_A\colon X \to \, :which for a given subset ''A'' of ''X'', has value 1 at point ...
of the Wishart distribution is :\Theta \mapsto \operatorname\left \, \exp\left( \,i \operatorname\left(\,\mathbf\,\right)\,\right)\, \right= \left, \, 1 - 2i\, \,\, \^ where denotes expectation. (Here is any matrix with the same dimensions as , indicates the identity matrix, and is a square root of ). Properly interpreting this formula requires a little care, because noninteger complex powers are multivalued; when is noninteger, the correct branch must be determined via
analytic continuation In complex analysis, a branch of mathematics, analytic continuation is a technique to extend the domain of definition of a given analytic function. Analytic continuation often succeeds in defining further values of a function, for example in a n ...
.


Theorem

If a random matrix has a Wishart distribution with degrees of freedom and variance matrix — write \mathbf\sim\mathcal_p(,m) — and is a matrix of rank , then :\mathbf\mathbf^T \sim \mathcal_q\left(^T,m\right).


Corollary 1

If is a nonzero constant vector, then: :\sigma_z^ \, ^T\mathbf \sim \chi_m^2. In this case, \chi_m^2 is the
chi-squared distribution In probability theory and statistics, the chi-squared distribution (also chi-square or \chi^2-distribution) with k degrees of freedom is the distribution of a sum of the squares of k independent standard normal random variables. The chi-squar ...
and \sigma_z^2=^T (note that \sigma_z^2 is a constant; it is positive because is positive definite).


Corollary 2

Consider the case where (that is, the -th element is one and all others zero). Then corollary 1 above shows that :\sigma_^ \, w_\sim \chi^2_m gives the marginal distribution of each of the elements on the matrix's diagonal.
George Seber George Arthur Frederick Seber (born 6 April 1938) is an Australian-born New Zealand statistician. Since his retirement from academic life, he has worked as a counsellor. Biography Born in Sydney in 1938, Seber emigrated to New Zealand with his ...
points out that the Wishart distribution is not called the “multivariate chi-squared distribution” because the marginal distribution of the off-diagonal elements is not chi-squared. Seber prefers to reserve the term multivariate for the case when all univariate marginals belong to the same family.


Estimator of the multivariate normal distribution

The Wishart distribution is the
sampling distribution In statistics, a sampling distribution or finite-sample distribution is the probability distribution of a given random-sample-based statistic. If an arbitrarily large number of samples, each involving multiple observations (data points), were se ...
of the maximum-likelihood estimator (MLE) of the covariance matrix of a
multivariate normal distribution In probability theory and statistics, the multivariate normal distribution, multivariate Gaussian distribution, or joint normal distribution is a generalization of the one-dimensional ( univariate) normal distribution to higher dimensions. One ...
. A derivation of the MLE uses the
spectral theorem In mathematics, particularly linear algebra and functional analysis, a spectral theorem is a result about when a linear operator or matrix can be diagonalized (that is, represented as a diagonal matrix in some basis). This is extremely useful b ...
.


Bartlett decomposition

The Bartlett decomposition of a matrix from a -variate Wishart distribution with scale matrix and degrees of freedom is the factorization: :\mathbf = ^T^T, where is the Cholesky factor of , and: :\mathbf A = \begin c_1 & 0 & 0 & \cdots & 0\\ n_ & c_2 &0 & \cdots& 0 \\ n_ & n_ & c_3 & \cdots & 0\\ \vdots & \vdots & \vdots &\ddots & \vdots \\ n_ & n_ & n_ &\cdots & c_p \end where c_i^2 \sim \chi^2_ and independently. This provides a useful method for obtaining random samples from a Wishart distribution.


Marginal distribution of matrix elements

Let be a variance matrix characterized by
correlation coefficient A correlation coefficient is a numerical measure of some type of correlation, meaning a statistical relationship between two variables. The variables may be two columns of a given data set of observations, often called a sample, or two componen ...
and its lower Cholesky factor: :\mathbf = \begin \sigma_1^2 & \rho \sigma_1 \sigma_2 \\ \rho \sigma_1 \sigma_2 & \sigma_2^2 \end, \qquad \mathbf = \begin \sigma_1 & 0 \\ \rho \sigma_2 & \sqrt \sigma_2 \end Multiplying through the Bartlett decomposition above, we find that a random sample from the Wishart distribution is :\mathbf = \begin \sigma_1^2 c_1^2 & \sigma_1 \sigma_2 \left (\rho c_1^2 + \sqrt c_1 n_ \right ) \\ \sigma_1 \sigma_2 \left (\rho c_1^2 + \sqrt c_1 n_ \right ) & \sigma_2^2 \left(\left (1-\rho^2 \right ) c_2^2 + \left (\sqrt n_ + \rho c_1 \right )^2 \right) \end The diagonal elements, most evidently in the first element, follow the distribution with degrees of freedom (scaled by ) as expected. The off-diagonal element is less familiar but can be identified as a
normal variance-mean mixture In probability theory and statistics, a normal variance-mean mixture with mixing probability density g is the continuous probability distribution of a random variable Y of the form :Y=\alpha + \beta V+\sigma \sqrtX, where \alpha, \beta and \sigma ...
where the mixing density is a distribution. The corresponding marginal probability density for the off-diagonal element is therefore the variance-gamma distribution :f(x_) = \frac \cdot K_ \left(\frac\right) \exp where is the modified Bessel function of the second kind. Similar results may be found for higher dimensions, but the interdependence of the off-diagonal correlations becomes increasingly complicated. It is also possible to write down the
moment-generating function In probability theory and statistics, the moment-generating function of a real-valued random variable is an alternative specification of its probability distribution. Thus, it provides the basis of an alternative route to analytical results compar ...
even in the ''noncentral'' case (essentially the ''n''th power of Craig (1936) equation 10) although the probability density becomes an infinite sum of Bessel functions.


The range of the shape parameter

It can be shown that the Wishart distribution can be defined if and only if the shape parameter belongs to the set :\Lambda_p:=\\cup \left(p-1,\infty\right). This set is named after Gindikin, who introduced it in the 1970s in the context of gamma distributions on homogeneous cones. However, for the new parameters in the discrete spectrum of the Gindikin ensemble, namely, :\Lambda_p^*:=\, the corresponding Wishart distribution has no Lebesgue density.


Relationships to other distributions

* The Wishart distribution is related to the inverse-Wishart distribution, denoted by W_p^, as follows: If and if we do the change of variables , then \mathbf\sim W_p^(\mathbf^,n). This relationship may be derived by noting that the absolute value of the Jacobian determinant of this change of variables is , see for example equation (15.15) in. * In
Bayesian statistics Bayesian statistics is a theory in the field of statistics based on the Bayesian interpretation of probability where probability expresses a ''degree of belief'' in an event. The degree of belief may be based on prior knowledge about the event, ...
, the Wishart distribution is a
conjugate prior In Bayesian probability theory, if the posterior distribution p(\theta \mid x) is in the same probability distribution family as the prior probability distribution p(\theta), the prior and posterior are then called conjugate distributions, and t ...
for the precision parameter of the
multivariate normal distribution In probability theory and statistics, the multivariate normal distribution, multivariate Gaussian distribution, or joint normal distribution is a generalization of the one-dimensional ( univariate) normal distribution to higher dimensions. One ...
, when the mean parameter is known. * A generalization is the
multivariate gamma distribution In statistics, a matrix gamma distribution is a generalization of the gamma distribution to positive-definite matrices.Iranmanesh, Anis, M. Arashib and S. M. M. Tabatabaey (2010)"On Conditional Applications of Matrix Variate Normal Distribution" ' ...
. * A different type of generalization is the
normal-Wishart distribution In probability theory and statistics, the normal-Wishart distribution (or Gaussian-Wishart distribution) is a multivariate four-parameter family of continuous probability distributions. It is the conjugate prior of a multivariate normal distributio ...
, essentially the product of a
multivariate normal distribution In probability theory and statistics, the multivariate normal distribution, multivariate Gaussian distribution, or joint normal distribution is a generalization of the one-dimensional ( univariate) normal distribution to higher dimensions. One ...
with a Wishart distribution.


See also

*
Chi-squared distribution In probability theory and statistics, the chi-squared distribution (also chi-square or \chi^2-distribution) with k degrees of freedom is the distribution of a sum of the squares of k independent standard normal random variables. The chi-squar ...
* Complex Wishart distribution *
F-distribution In probability theory and statistics, the ''F''-distribution or F-ratio, also known as Snedecor's ''F'' distribution or the Fisher–Snedecor distribution (after Ronald Fisher and George W. Snedecor) is a continuous probability distribution ...
*
Gamma distribution In probability theory and statistics, the gamma distribution is a two- parameter family of continuous probability distributions. The exponential distribution, Erlang distribution, and chi-square distribution are special cases of the gamma dis ...
* Hotelling's T-squared distribution * Inverse-Wishart distribution *
Multivariate gamma distribution In statistics, a matrix gamma distribution is a generalization of the gamma distribution to positive-definite matrices.Iranmanesh, Anis, M. Arashib and S. M. M. Tabatabaey (2010)"On Conditional Applications of Matrix Variate Normal Distribution" ' ...
* Student's t-distribution *
Wilks' lambda distribution In statistics, Wilks' lambda distribution (named for Samuel S. Wilks), is a probability distribution used in multivariate hypothesis testing, especially with regard to the likelihood-ratio test and multivariate analysis of variance (MANOVA). Defin ...


References


External links


A C++ library for random matrix generator
{{DEFAULTSORT:Wishart Distribution Continuous distributions Multivariate continuous distributions Covariance and correlation Random matrices Conjugate prior distributions Exponential family distributions