The Kaiser–Meyer–Olkin (KMO) test is a statistical measure to determine how suited data is for
factor analysis
Factor analysis is a statistical method used to describe variability among observed, correlated variables in terms of a potentially lower number of unobserved variables called factors. For example, it is possible that variations in six observed ...
. The test measures sampling adequacy for each variable in the model and the complete model. The statistic is a measure of the proportion of variance among variables that might be common variance. The higher the proportion, the higher the KMO-value, the more suited the data is to factor analysis.
History
Henry Kaiser introduced a Measure of Sampling Adequacy (MSA) of factor analytic data matrices in 1970.
Kaiser and Rice then modified it in 1974.
Measure of sampling adequacy
The measure of sampling adequacy is calculated for each indicator as
:
and indicates to what extent an indicator is suitable for a factor analysis.
Kaiser–Meyer–Olkin criterion
: The Kaiser–Meyer–Olkin criterion is calculated and returns values between 0 and 1.
:
:
Here
is the correlation between the variable in question and another, and
is the partial correlation.
This is a function of the squared elements of the `image' matrix compared to the squares of the original correlations. The overall MSA as well as estimates for each item are found. The index is known as the Kaiser–Meyer–Olkin (KMO) index.
Interpretation of result
In flamboyant fashion, Kaiser proposed that a KMO > 0.9 was marvelous, in the 0.80s, meritorious, in the 0.70s, middling, in the 0.60s, mediocre, in the 0.50s, miserable, and less than 0.5 would be unacceptable.
In general, KMO values between 0.8 and 1 indicate the sampling is adequate. KMO values less than 0.6 indicate the sampling is not adequate and that remedial action should be taken. In contrast, others set this cutoff value at 0.5.
A KMO value close to zero means that there are large partial correlations compared to the sum of correlations. In other words, there are widespread correlations which would be a large problem for factor analysis.
An alternative measure of whether a matrix is factorable is the
Bartlett test
In statistics, Bartlett's test, named after Maurice Stevenson Bartlett, is used to test homoscedasticity, that is, if multiple samples are from populations with equal variances. Some statistical tests, such as the analysis of variance, assume tha ...
, which tests the degree that the matrix deviates from an
identity matrix
In linear algebra, the identity matrix of size n is the n\times n square matrix with ones on the main diagonal and zeros elsewhere.
Terminology and notation
The identity matrix is often denoted by I_n, or simply by I if the size is immaterial ...
.
Example in R
If the following is run in
R with the library(psych)
library(psych)
set.seed(5L)
five.samples <- data.frame("A"=rnorm(100), "B"=rnorm(100), "C"=rnorm(100),
"D"=rnorm(100), "E"=rnorm(100))
cor(five.samples)
KMO(five.samples)
The following is produced:
Kaiser-Meyer-Olkin factor adequacy
Call: KMO(r = five.samples)
Overall MSA = 0.53
MSA for each item =
A B C D E
0.52 0.56 0.52 0.48 0.54
This shows that the data is not that suited to Factor Analysis.
See also
*
Box's M test
Box's ''M'' test is a multivariate statistical test used to check the equality of multiple variance-covariance matrices. The test is commonly used to test the assumption of homogeneity of variances and covariances in MANOVA and linear discrimin ...
*
Levene's test In statistics, Levene's test is an inferential statistic used to assess the equality of variances for a variable calculated for two or more groups. Some common statistical procedures assume that variances of the populations from which different sam ...
*
Bartlett's test
In statistics, Bartlett's test, named after Maurice Stevenson Bartlett, is used to test homoscedasticity, that is, if multiple samples are from populations with equal variances. Some statistical tests, such as the analysis of variance, assume th ...
References
{{DEFAULTSORT:Kaiser-Meyer-Olkin test
Analysis of variance
Statistical tests