Kaiser–Meyer–Olkin Test
   HOME

TheInfoList



OR:

The Kaiser–Meyer–Olkin (KMO) test is a statistical measure to determine how suited data is for
factor analysis Factor analysis is a statistical method used to describe variability among observed, correlated variables in terms of a potentially lower number of unobserved variables called factors. For example, it is possible that variations in six observed ...
. The test measures sampling adequacy for each variable in the model and the complete model. The statistic is a measure of the proportion of variance among variables that might be common variance. The higher the proportion, the higher the KMO-value, the more suited the data is to factor analysis.


History

Henry Kaiser introduced a Measure of Sampling Adequacy (MSA) of factor analytic data matrices in 1970. Kaiser and Rice then modified it in 1974.


Measure of sampling adequacy

The measure of sampling adequacy is calculated for each indicator as :MSA_j = \frac and indicates to what extent an indicator is suitable for a factor analysis.


Kaiser–Meyer–Olkin criterion

: The Kaiser–Meyer–Olkin criterion is calculated and returns values between 0 and 1. : :KMO = \frac Here r_ is the correlation between the variable in question and another, and p_ is the partial correlation. This is a function of the squared elements of the `image' matrix compared to the squares of the original correlations. The overall MSA as well as estimates for each item are found. The index is known as the Kaiser–Meyer–Olkin (KMO) index.


Interpretation of result

In flamboyant fashion, Kaiser proposed that a KMO > 0.9 was marvelous, in the 0.80s, meritorious, in the 0.70s, middling, in the 0.60s, mediocre, in the 0.50s, miserable, and less than 0.5 would be unacceptable. In general, KMO values between 0.8 and 1 indicate the sampling is adequate. KMO values less than 0.6 indicate the sampling is not adequate and that remedial action should be taken. In contrast, others set this cutoff value at 0.5. A KMO value close to zero means that there are large partial correlations compared to the sum of correlations. In other words, there are widespread correlations which would be a large problem for factor analysis. An alternative measure of whether a matrix is factorable is the Bartlett test, which tests the degree that the matrix deviates from an identity matrix.


Example in R

If the following is run in R with the library(psych) library(psych) set.seed(5L) five.samples <- data.frame("A"=rnorm(100), "B"=rnorm(100), "C"=rnorm(100),                     "D"=rnorm(100), "E"=rnorm(100)) cor(five.samples) KMO(five.samples) The following is produced: Kaiser-Meyer-Olkin factor adequacy Call: KMO(r = five.samples) Overall MSA = 0.53 MSA for each item = A B C D E 0.52 0.56 0.52 0.48 0.54 This shows that the data is not that suited to Factor Analysis.


See also

*
Box's M test Box's ''M'' test is a multivariate statistical test used to check the equality of multiple variance-covariance matrices. The test is commonly used to test the assumption of homogeneity of variances and covariances in MANOVA and linear discrimina ...
*
Levene's test In statistics, Levene's test is an inferential statistic used to assess the equality of variances for a variable calculated for two or more groups. Some common statistical procedures assume that variances of the populations from which different sam ...
*
Bartlett's test In statistics, Bartlett's test, named after Maurice Stevenson Bartlett, is used to test homoscedasticity, that is, if multiple samples are from populations with equal variances. Some statistical tests, such as the analysis of variance, assume tha ...


References

{{DEFAULTSORT:Kaiser-Meyer-Olkin test Analysis of variance Statistical tests