In
probability theory
Probability theory or probability calculus is the branch of mathematics concerned with probability. Although there are several different probability interpretations, probability theory treats the concept in a rigorous mathematical manner by expre ...
and
statistics
Statistics (from German language, German: ', "description of a State (polity), state, a country") is the discipline that concerns the collection, organization, analysis, interpretation, and presentation of data. In applying statistics to a s ...
, the noncentral chi-squared distribution (or noncentral chi-square distribution, noncentral
distribution) is a
noncentral generalization of the
chi-squared distribution. It often arises in the
power analysis of statistical tests in which the null distribution is (perhaps asymptotically) a chi-squared distribution; important examples of such tests are the
likelihood-ratio tests.
Definitions
Background
Let
be ''k''
independent,
normally distributed random variables with means
and unit variances. Then the random variable
:
is distributed according to the noncentral chi-squared distribution. It has two parameters:
which specifies the number of
degrees of freedom
In many scientific fields, the degrees of freedom of a system is the number of parameters of the system that may vary independently. For example, a point in the plane has two degrees of freedom for translation: its two coordinates; a non-infinite ...
(i.e. the number of
), and
which is related to the mean of the random variables
by:
:
is sometimes called the
noncentrality parameter. Note that some references define
in other ways, such as half of the above sum, or its square root.
This distribution arises in
multivariate statistics as a derivative of the
multivariate normal distribution. While the central
chi-squared distribution is the squared
norm of a
random vector with
distribution (i.e., the squared distance from the origin to a point taken at random from that distribution), the non-central
is the squared norm of a random vector with
distribution. Here
is a zero vector of length ''k'',
and
is the
identity matrix
In linear algebra, the identity matrix of size n is the n\times n square matrix with ones on the main diagonal and zeros elsewhere. It has unique properties, for example when the identity matrix represents a geometric transformation, the obje ...
of size ''k''.
Density
The
probability density function
In probability theory, a probability density function (PDF), density function, or density of an absolutely continuous random variable, is a Function (mathematics), function whose value at any given sample (or point) in the sample space (the s ...
(pdf) is given by
:
where
is distributed as chi-squared with
degrees of freedom.
From this representation, the noncentral chi-squared distribution is seen to be a Poisson-weighted
mixture
In chemistry, a mixture is a material made up of two or more different chemical substances which can be separated by physical method. It is an impure substance made up of 2 or more elements or compounds mechanically mixed together in any proporti ...
of central chi-squared distributions. Suppose that a random variable ''J'' has a
Poisson distribution with mean
, and the
conditional distribution of ''Z'' given ''J'' = ''i'' is chi-squared with ''k'' + 2''i'' degrees of freedom. Then the
unconditional distribution of ''Z'' is non-central chi-squared with ''k'' degrees of freedom, and non-centrality parameter
.
Alternatively, the pdf can be written as
:
where
is a modified
Bessel function of the first kind given by
:
Using the relation between
Bessel functions and
hypergeometric functions, the pdf can also be written as:
:
The case ''k'' = 0 (
zero degrees of freedom), in which case the distribution has a discrete component at zero, is discussed by Torgersen (1972) and further by Siegel (1979).
Derivation of the pdf
The derivation of the probability density function is most easily done by performing the following steps:
# Since
have unit variances, their joint distribution is spherically symmetric, up to a location shift.
# The spherical symmetry then implies that the distribution of
depends on the means only through the squared length,
. Without loss of generality, we can therefore take
and
.
# Now derive the density of
(i.e. the ''k'' = 1 case). Simple transformation of random variables shows that
:::
::where
is the standard normal density.
# Expand the
cosh term in a
Taylor series. This gives the Poisson-weighted mixture representation of the density, still for ''k'' = 1. The indices on the chi-squared random variables in the series above are 1 + 2''i'' in this case.
# Finally, for the general case. We've assumed, without loss of generality, that
are standard normal, and so
has a ''central'' chi-squared distribution with (''k'' − 1) degrees of freedom, independent of
. Using the poisson-weighted mixture representation for
, and the fact that the sum of chi-squared random variables is also a chi-square, completes the result. The indices in the series are (1 + 2''i'') + (''k'' − 1) = ''k'' + 2''i'' as required.
Properties
Moment generating function
The
moment-generating function is given by
:
Moments
The first few raw
moments are:
:
:
:
:
The first few central
moments are:
:
:
:
The ''n''th
cumulant is
:
Hence
:
Cumulative distribution function
Again using the relation between the central and noncentral chi-squared distributions, the
cumulative distribution function
In probability theory and statistics, the cumulative distribution function (CDF) of a real-valued random variable X, or just distribution function of X, evaluated at x, is the probability that X will take a value less than or equal to x.
Ever ...
(cdf) can be written as
:
where
is the cumulative distribution function of the central chi-squared distribution with ''k'' degrees of freedom which is given by
:
:and where
is the
lower incomplete gamma function.
The
Marcum Q-function can also be used to represent the cdf.
:
When the degrees of freedom ''k'' is positive odd integer, we have a closed form expression for the complementary cumulative distribution function given by
[A. Annamalai, C. Tellambura and John Matyjas (2009). "A New Twist on the Generalized Marcum Q-Function ''Q''''M''(''a'', ''b'') with Fractional-Order ''M'' and its Applications". ''2009 6th IEEE Consumer Communications and Networking Conference'', 1–5, ]
:
where ''n'' is non-negative integer, ''Q'' is the
Gaussian Q-function, and ''I'' is the modified Bessel function of first kind with half-integer order. The modified Bessel function of first kind with half-integer order in itself can be represented as a finite sum in terms of
hyperbolic functions.
In particular, for ''k'' = 1, we have
:
Also, for ''k'' = 3, we have
:
Approximation (including for quantiles)
Abdel-Aty derives (as "first approx.") a non-central
Wilson–Hilferty transformation:
is approximately
normally distributed,
i.e.,
:
which is quite accurate and well adapting to the noncentrality. Also,
becomes
for
, the
(central) chi-squared case.
Sankaran discusses a number of
closed form approximations for the
cumulative distribution function
In probability theory and statistics, the cumulative distribution function (CDF) of a real-valued random variable X, or just distribution function of X, evaluated at x, is the probability that X will take a value less than or equal to x.
Ever ...
. In an earlier paper, he derived and states the following approximation:
:
where
:
denotes the
cumulative distribution function
In probability theory and statistics, the cumulative distribution function (CDF) of a real-valued random variable X, or just distribution function of X, evaluated at x, is the probability that X will take a value less than or equal to x.
Ever ...
of the
standard normal distribution;
:
:
:
This and other approximations are discussed in a later text book.
More recently, since the CDF of non-central chi-squared distribution with odd degree of freedom can be exactly computed, the CDF for even degree of freedom can be approximated by exploiting the monotonicity and log-concavity properties of Marcum-Q function as
:
Another approximation that also serves as an upper bound is given by
:
For a given probability, these formulas are easily inverted to provide the corresponding approximation for
, to compute approximate quantiles.
Related distributions
*If
is
chi-square distributed,
, then
is also non-central chi-square distributed:
*A linear combination of independent noncentral chi-squared variables
, is
generalized chi-square distributed.
*If
and
and
is independent of
then a
noncentral ''F''-distributed variable is developed as
*If
, then
*If
, then
takes the
Rice distribution with parameter
.
*Normal approximation: if
, then
in distribution as either
or
.
*If
and
, where
are independent, then
.
*In general, for an independent finite set of
,
, the sum of these non-central chi-square distributed random variables
has the distribution
where
,
. This can be seen using moment generating functions as follows:
by the independence of the
random variables. It remains to plug in the MGF for the non-central chi square distributions into the product and compute the new MGF – this is left as an exercise. Alternatively it can be seen via the interpretation in the background section above as sums of squares of independent normally distributed random variables with variances of 1 and the specified means.
* The ''complex noncentral chi-squared distribution'' has applications in radio communication and radar systems. Let
be independent scalar
complex random variable
In probability theory and statistics, complex random variables are a generalization of real-valued random variables to complex numbers, i.e. the possible values a complex random variable may take are complex numbers. Complex random variables can al ...
s with noncentral circular symmetry, means of
and unit variances:
. Then the real random variable
is distributed according to the complex noncentral chi-squared distribution, which is effectively a scaled (by 1/2) non-central
with twice the degrees of freedom and twice the noncentrality parameter:
::
,
:where
.
Transformations
Sankaran (1963) discusses the transformations of the form
. He analyzes the expansions of the
cumulants of
up to the term
and shows that the following choices of
produce reasonable results:
*
makes the second cumulant of
approximately independent of
*
makes the third cumulant of
approximately independent of
*
makes the fourth cumulant of
approximately independent of
Also, a simpler transformation
can be used as a
variance stabilizing transformation that produces a random variable with mean
and variance
.
Usability of these transformations may be hampered by the need to take the square roots of negative numbers.
Occurrence and applications
Use in tolerance intervals
Two-sided normal
regression tolerance intervals can be obtained based on the noncentral chi-squared distribution.
[, p. 32] This enables the calculation of a statistical interval within which, with some confidence level, a specified proportion of a sampled population falls.
Notes
References
* Abramowitz, M. and Stegun, I. A. (1972), ''
Handbook of Mathematical Functions'', Dover.
* Johnson, N. L., Kotz, S., Balakrishnan, N. (1995), ''Continuous Univariate Distributions, Volume 2 (2nd Edition)'', Wiley.
* Muirhead, R. (2005) ''Aspects of Multivariate Statistical Theory'' (2nd Edition). Wiley.
*
{{ProbDistributions, continuous-semi-infinite
Continuous distributions
c