Factor Analysis Exercises
Factor Analysis Exercises
You are an expert in labor policies working for the European Commission and you have been asked
by your supervisor to compile a report on the situation of labor and social protection within the 27
countries belonging to the European Union. For this purpose, you decide to use data made
available from the World Bank for the year 2012.
All variables consist of percentages, therefore all of them are expressed on the same scale.
1) Based on Table 4.1, which problem do you have to deal with in order to proceed with the
analysis?
I notice that there are some missing values. I could therefore opt either for the removal
of the countries with missing information, either listwise of pairwise, or for replacing the
missing values with the mean of the variable.
2) What does the Bartlett’s test of sphericity tell you about the correlation between the
items? What is the null hypothesis behind the test? Given a value a value of 673.178 (p-
value=.000) for Bartlett’s test, would you proceed with the analysis or not? Are you
confident with that or would you need some other piece of information?
Bartlett’s test of sphericity is a global test on the correlation matrix, whose null
hypothesis states that the correlation between the variables is equal to zero, thus the
correlation matrix is an identity matrix. Just considering the result of the test is not
enough to proceed with the analysis, since the test just tells me that there is some
correlation between the variables, but I do not know how strong it is and among how
many variables. For this information, I need the value of the KMO measure.
3) You decide to remove two variables from the analysis (“employment in industry” and
“employers”). Looking at the main diagonal of the new anti-image matrix, you see that the
values range from 0.644 (“Unemployment, total”) to 0.800 (“Vulnerable employment,
total”). What do these values tell us? Are you confident enough with going on with the
analysis (the new KMO measure is 0.742)?
The anti-image matrix reports the KMO MSA measures for each of the variables, that is to
say, the amount of common variance between each variable and the remaining ones. The
smallest value, 0.644, indicates that the inter-correlation between the item and the other
ones is just mediocre. However, if the other values are above 0.7, the degree of inter-
correlation should be satisfactory. Moreover, the KMO MSA is larger than 0.7, indicating that
the total degree of inter-correlation in the data is satisfactory. Hence, I would proceed with
the analysis.
4) What are communalities? What is the difference between the initial communalities for a
“principal axis factoring analysis” and the initial communalities for a “principal component
analysis”?
The communality of a variable is the proportion of each variable’s variance that is in
common with the other variable and therefore can be explained by the factors. It can be
defined as the sum of squared factor loadings for the variables.
The initial communalities under the principal axis factoring analysis aim to describe the
share of common variance between each variable and all the remaining ones, thus they
are numbers between 0 and 1. This happens because this method is not interested in
explaining the whole variance in the data, but only the common variance between the
variables.
Instead, the initial communalities under the PCA are all equal to one because now the
analysis aims to reduce the data dimension having as goal to explain as much as possible
of the whole variability in the data, not only the one in common between the variables.
5) Table 4.3 shows the total variance explained by each of the factors. What possible criteria
might have been used to select the number of factors? Is it OK if the two factors explain
only 81% of the whole variance?
One possible criterion is to select those factors characterized by an initial eigenvalue
bigger than one, which indicates that the factor accounts for more variability than a
single variable. Another possible criterion is to look at the scree plot, which plots the
eigenvalues against the factor number. The optimal number of factors is the one that is
followed by an almost flat line. This indicates that each successive factor accounts for
increasingly smaller amounts of the total variance.
I am fine with the fact that only around 80% of total variance is explained by the factors.
The aim of FA is to reduce the dimension of data to a smaller set of variables keeping as
large as possible their capacity to explain most of the total variance in the initial data. In
this case, the result is very good, because 80% of the total variance is explained by two
factors, compared to the ten initial variables.
6) If you asked for rotation, SPSS output finally shows you two matrices, one with the
unrotated factors and the other with the rotated ones. Which one would you consider to
facilitate your task of interpreting the latent factors that you found? And why?
I would consider the matrix with the rotated factors. Indeed, rotation is a mathematical
transformation that doesn’t alter the factors and the quantity of variability that they
explain. Rather, by distributing the variance that they explain more evenly among the
variables, rotation makes interpretation of the factors more straightforward. Different
kinds of rotation are possible. For instance, varimax rotation gives totally uncorrelated
factors, while oblique rotation gives correlated factors.
7) Using Table 4.4, with the rotated factors, obtained through Varimax method with Kaiser
normalization, would you be able to compute the communality for the item “Long-term
unemployment”?
The communality for the item is given by
(0.097)^2+(0.912)^2 = 0.8412.
8) Instead of the Varimax method for rotation, say, you opted for the Direct Oblimin method.
What is the assumption behind this rotation method? You therefore got the output shown
in Table 4.5. Given this piece of information, would you be confident enough in using your
two new factors as independent variables in a regression model?
Direct Oblimin is a method for oblique rotation. This is not an orthogonal rotation of the
factors, therefore the factors are no longer uncorrelated. In this particular case, since the
correlation between the two factors is rather small, just 0.28, I would be confident
enough in using the two factors as independent variables in a regression model.
Factor 1 2
1 1.000 .282
2 .282 1.000