Multivariate Behrens–Fisher Problem
   HOME

TheInfoList



OR:

In
statistics Statistics (from German language, German: ''wikt:Statistik#German, Statistik'', "description of a State (polity), state, a country") is the discipline that concerns the collection, organization, analysis, interpretation, and presentation of ...
, the multivariate Behrens–Fisher problem is the problem of testing for the equality of means from two
multivariate normal In probability theory and statistics, the multivariate normal distribution, multivariate Gaussian distribution, or joint normal distribution is a generalization of the one-dimensional (univariate) normal distribution to higher dimensions. One d ...
distributions when the covariance matrices are unknown and possibly not equal. Since this is a generalization of the univariate Behrens-Fisher problem, it inherits all of the difficulties that arise in the univariate problem.


Notation and problem formulation

Let X_ \sim \mathcal_p(\mu_i,\, \Sigma_i) \ \ (j=1,\dots,n_i; \ \ i=1,2)\ be independent random samples from two p-variate normal distributions with unknown mean vectors \mu_i and unknown dispersion matrices \Sigma_i. The index i refers to the first or second population, and the jth observation from the ith population is X_. The multivariate Behrens–Fisher problem is to test the null hypothesis H_0 that the means are equal versus the alternative H_1 of non-equality: : H_0 : \mu_1 = \mu_2 \ \ \text \ \ H_1 : \mu_1 \neq \mu_2. Define some statistics, which are used in the various attempts to solve the multivariate Behrens–Fisher problem, by : \begin \bar &= \frac \sum_^ X_, \\ A_i &= \sum_^ (X_ - \bar)(X_ - \bar)', \\ S_i &= \frac A_i, \\ \tilde &= \fracS_i, \\ \tilde &= \tilde + \tilde, \quad \text \\ T^2 & = (\bar - \bar)'\tilde^(\bar - \bar). \end The sample means \bar and sum-of-squares matrices A_i are
sufficient In logic and mathematics, necessity and sufficiency are terms used to describe a conditional or implicational relationship between two statements. For example, in the conditional statement: "If then ", is necessary for , because the truth of ...
for the multivariate normal parameters \mu_i, \Sigma_i,\ (i=1,2), so it suffices to perform inference be based on just these statistics. The distributions of \bar and A_i are independent and are, respectively,
multivariate normal In probability theory and statistics, the multivariate normal distribution, multivariate Gaussian distribution, or joint normal distribution is a generalization of the one-dimensional (univariate) normal distribution to higher dimensions. One d ...
and Wishart: : \begin \bar &\sim \mathcal_p \left(\mu_i, \Sigma_i/n_i \right), \\ A_i &\sim W_p(\Sigma_i, n_i - 1). \end


Background

In the case where the dispersion matrices are equal, the distribution of the T^2 statistic is known to be an
F distribution In probability theory and statistics, the ''F''-distribution or F-ratio, also known as Snedecor's ''F'' distribution or the Fisher–Snedecor distribution (after Ronald Fisher and George W. Snedecor) is a continuous probability distribution t ...
under the null and a
noncentral F-distribution In probability theory and statistics, the noncentral ''F''-distribution is a continuous probability distribution that is a noncentral distribution, noncentral generalization of the (ordinary) F-distribution, ''F''-distribution. It describes the di ...
under the alternative. The main problem is that when the true values of the dispersion matrix are unknown, then under the null hypothesis the probability of rejecting H_0 via a T^2 test depends on the unknown dispersion matrices. In practice, this dependency harms inference when the dispersion matrices are far from each other or when the sample size is not large enough to estimate them accurately. Now, the mean vectors are independently and normally distributed, : \bar \sim \mathcal_p \left(\mu_i, \Sigma_i/n_i \right), but the sum A_1 + A_2 does not follow the Wishart distribution, which makes inference more difficult.


Proposed solutions

Proposed solutions are based on a few main strategies: * Compute statistics which mimick the T^2 statistic and which have an approximate F distribution with estimated degrees of freedom (df). * Use generalized p-values based on generalized test variables. * Use Roy's union-intersection principle


Approaches using the ''T''2 with approximate degrees of freedom

Below, \mathrm indicates the
trace operator In mathematics, the trace operator extends the notion of the restriction of a function to the boundary of its domain to "generalized" functions in a Sobolev space. This is particularly important for the study of partial differential equations with ...
.


Yao (1965)

(as cited by ) : T^2 \sim \fracF_, where : \begin \nu &= \left \frac \left( \frac \right)^2 + \frac \left( \frac \right)^ \right, \\ \bar_d & = \bar_-\bar_2. \end


Johansen (1980)

(as cited by ) : T^2 \sim q F_, where : \begin q &= p + 2D - \frac, \\ \nu &= \frac, \\ \end and : \begin D = \frac\sum_^2 \frac \Bigg\. \\ \end


Nel and Van der Merwe's (1986)

(as cited by ) : T^2 \sim \fracF_, where : \nu = \frac .


Comments on performance

Kim (1992) proposed a solution that is based on a variant of T^2. Although its power is high, the fact that it is not invariant makes it less attractive. Simulation studies by Subramaniam and Subramaniam (1973) show that the size of Yao's test is closer to the nominal level than that of James's. Christensen and Rencher (1997) performed numerical studies comparing several of these testing procedures and concluded that Kim and Nel and Van der Merwe's tests had the highest power. However, these two procedures are not invariant.


Krishnamoorthy and Yu (2004)

Krishnamoorthy and Yu (2004) proposed a procedure which adjusts in Nel and Var der Merwe (1986)'s approximate df for the denominator of T^2 under the null distribution to make it invariant. They show that the approximate degrees of freedom lies in the interval \left min\,n_1+n_2-2\right/math> to ensure that the degrees of freedom is not negative. They report numerical studies that indicate that their procedure is as powerful as Nel and Van der Merwe's test for smaller dimension, and more powerful for larger dimension. Overall, they claim that their procedure is the better than the invariant procedures of Yao (1965) and Johansen (1980). Therefore, Krishnamoorthy and Yu's (2004) procedure has the best known size and power as of 2004. The test statistic T^2 in Krishnmoorthy and Yu's procedure follows the distribution T^2 \sim \nu pF_/(\nu-p+1), where : \nu = \frac .


References

* Rodríguez-Cortés, F. J. and Nagar, D. K. (2007). Percentage points for testing equality of mean vectors. ''Journal of the Nigerian Mathematical Society'', 26:85–95. *Gupta, A. K., Nagar, D. K., Mateu, J. and Rodríguez-Cortés, F. J. (2013). Percentage points of a test statistic useful in manova with structured covariance matrices. ''Journal of Applied Statistical Science'', 20:29-41. {{DEFAULTSORT:Multivariate Behrens-Fisher problem Multivariate continuous distributions Normal distribution