Lecture 10
Lecture 10
Factor Analysis
𝑋
⏟ = ⏟
𝜇 + ⏟
𝐿 𝐹
⏟ + ⏟
𝜀
(𝑝×1) (𝑝×1) (𝑝×𝑚) (𝑚×1) (𝑝×1)
𝜇𝑖 = 𝑚𝑒𝑎𝑛 𝑜𝑓 𝑣𝑎𝑟𝑖𝑎𝑏𝑙𝑒 𝑖
𝜀𝑖 = 𝑖𝑡ℎ 𝑠𝑝𝑒𝑐𝑖𝑓𝑖𝑐 𝑓𝑎𝑐𝑡𝑜𝑟
𝐹𝑗 = 𝑗𝑡ℎ 𝑐𝑜𝑚𝑚𝑜𝑛 𝑓𝑎𝑐𝑡𝑜𝑟
𝑙𝑖𝑗 = 𝑙𝑜𝑎𝑑𝑖𝑛𝑔 𝑜𝑓 𝑡ℎ𝑒 𝑖𝑡ℎ 𝑣𝑎𝑟𝑖𝑎𝑏𝑙𝑒 𝑜𝑛 𝑡ℎ𝑒 𝑗𝑡ℎ 𝑓𝑎𝑐𝑡𝑜𝑟
The unobservable random vector F and 𝜺 satisfy the following conditions:
1. 𝐶𝑜𝑣(𝑋) = 𝐿𝐿′ + 𝜳
or
2 2
𝑉𝑎𝑟(𝑋𝑖 ) = 𝑙𝑖1 + ⋯ … … … . +𝑙𝑖𝑚 + 𝛹𝑖
𝐶𝑜𝑣(𝑋𝑖 , 𝑋𝑘 ) = 𝑙𝑖1 𝑙𝑘1 + ⋯ … … … . +𝑙𝑖𝑚 𝑙𝑘𝑚
2. 𝐶𝑜𝑣(𝑋, 𝐹) = 𝐿
or
𝐶𝑜𝑣(𝑋𝑖 , 𝐹𝑗 ) = 𝑙𝑖𝑗
➢ The portion of the variance of the ith variable contributed by the m common
factors is called the ith communality.
➢ The portion of 𝑽𝒂𝒓 (𝑿𝒊 ) = 𝝈𝒊𝒊 due to the specific factor is often called the
uniqueness, or specific variance.
➢ Denoting the ith communality by ℎ2 , we that
2 2 2
𝜎
⏟𝑖𝑖 =⏟
𝑙𝑖1 + 𝑙𝑖2 + ⋯ … … . +𝑙𝑖𝑚 + 𝛹
⏟𝑖
𝑉𝑎𝑟 (𝑋𝑖 ) = 𝑐𝑜𝑚𝑚𝑛𝑎𝑙𝑖𝑡𝑦 + 𝑠𝑝𝑒𝑐𝑖𝑓𝑖𝑐 𝑣𝑎𝑟𝑖𝑎𝑛𝑐𝑒
or
ℎ𝑖2 = 𝑙𝑖1
2 2
+ 𝑙𝑖2 2
+ ⋯ … . . +𝑙𝑖𝑚
and
𝜎𝑖𝑖 = ℎ2 + 𝛹𝑖 𝑖 = 1.2 … … . . , 𝑝
The ith communality is the sum of squares of the loadings of the ith variable on the
m common factor.
Example: (Verify that the relation ∑= 𝐿𝐿′ + 𝜳 for two factors) Consider the
covariance matrix
19 30 2 12
30 57 5 23
∑=[ ]
2 5 38 47
12 23 47 68
The equality
19 30 2 12 4 1 2 0 0 0
30 57 5 23 7 2 4 7 −1 1 0 4 0 0
[ ]=[ ][ ]+[ ]
2 5 38 47 −1 6 1 2 6 8 0 0 1 0
12 23 47 68 1 8 0 0 0 3
or
∑= 𝐿𝐿′ + 𝜳
𝑙11 𝑙12 4 1
𝑙 𝑙 7 2
𝐿 = [ 21 22 ] = [ ]
𝑙31 𝑙32 −1 6
𝑙41 𝑙42 1 8
𝜓1 0 0 0 2 0 0 0
0 𝜓2 0 0 0 4 0 0
𝜳=[ ]=[ ]
0 0 𝜓3 0 0 0 1 0
0 0 0 𝜓4 0 0 0 3
19
⏟ = ⏟2
4 + 12 + ⏟ 2
𝑣𝑎𝑟𝑖𝑛𝑎𝑐𝑒 𝑐𝑜𝑚𝑚𝑢𝑛𝑎𝑙𝑖𝑡𝑦 + 𝑠𝑝𝑒𝑐𝑖𝑓𝑖𝑐 𝑣𝑎𝑟𝑖𝑎𝑛𝑐𝑒
➢ The estimated specific variances are provided by the diagonal elements of the
̃′ , so
matrix 𝑆 − 𝐿𝐿
𝜓̃1 0 ⋯ 0
̃
𝛹̃ = 0 𝜓2 ⋯ 0 with 𝜓̃𝑖 = 𝑠𝑖𝑖 − ∑𝑚 ̃2
𝑖=1 𝑙𝑖𝑗 … … … . (2)
⋮ ⋮ ⋱ ⋮
[0 0 ⋯ 𝜓̃𝑝 ]
ℎ̃𝑖2 = 𝑙̃𝑖1
2
+ 𝑙̃𝑖2
2
+ ⋯ … … . +𝑙̃𝑖𝑚
2
… … … … . . (3)
➢ Ideally, the contributions of the first few factors to the sample variances of
the variables should be large.
✓ The contribution to the sample variance 𝑠𝑖𝑖 from the first common
factor is 𝑙̃𝑖1
2
. The contribution to the total sample variance,
In general,
𝜆̂𝑗
𝑃𝑟𝑜𝑝𝑜𝑟𝑡𝑖𝑜𝑛 𝑜𝑓 𝑡𝑜𝑡𝑎𝑙 𝑓𝑜𝑟 𝑎 𝑓𝑎𝑐𝑡𝑜𝑟 𝑎𝑛𝑎𝑙𝑦𝑠𝑖𝑠 𝑜𝑓 𝑆
𝑠11 + 𝑠22 + ⋯ + 𝑠𝑝𝑝
( 𝑠𝑎𝑚𝑝𝑙𝑒 𝑣𝑎𝑟𝑖𝑎𝑛𝑐𝑒 ) =
𝑑𝑢𝑒 𝑡𝑜 𝑗𝑡ℎ 𝑓𝑎𝑐𝑡𝑜𝑟 𝜆̂𝑗
𝑓𝑜𝑟 𝑎 𝑓𝑎𝑐𝑡𝑜𝑟 𝑎𝑛𝑎𝑙𝑦𝑠𝑖𝑠 𝑜𝑓 𝑅
{ 𝑝
Attribute (variable) 1 2 3 4 5
𝑇𝑎𝑠𝑡𝑒 1 1.00 . 02 . 96 . 42 . 01
𝐺𝑜𝑜𝑑 𝑏𝑢𝑦 𝑓𝑜𝑟 𝑚𝑜𝑛𝑒𝑦 2 . 02 1.00 . 13 . 71 . 85
𝐹𝑙𝑎𝑣𝑜𝑟 3 . 96 . 13 1.00 . 50 . 11
𝑆𝑢𝑖𝑡𝑎𝑏𝑙𝑒 𝑓𝑜𝑟 𝑠𝑛𝑎𝑐𝑘 4 . 42 . 71 . 50 1.00 . 79
𝑃𝑟𝑜𝑣𝑖𝑑𝑒𝑠 𝑙𝑜𝑡𝑠 𝑜𝑓 𝑒𝑛𝑒𝑟𝑔𝑦 5 [ . 01 . 85 . 11 . 79 1.00]
➢ It is clear from the shaded entries in the correlation matrix that variables 1 and
3 and variables 2 and 5 form groups.
➢ Variable 4 is "closer" to the (2, 5) group than the (1, 3) group.
➢ Given these results and the small number of variables, we might expect that
the apparent linear relationships between the variables can be explained in
terms of, at most, two or three common factors.
➢ The first two eigenvalues, 𝜆̂1 = 2.85 and 𝜆̂2 = 1.81, of R are the only
eigenvalues greater than unity.
➢ Moreover, m = 2 common factors will account for a cumulative proportion
Now
. 56 . 82
. 78 −.53
̃ +𝜳
𝐿𝐿′ ̃ = . 65 . 75 [. 56 . 78 . 65 . 94 . 80
]
. 82 −.53 . 75 −.10 −.54
. 94 −.10
[. 80 −.54]
. 02 0 0 0 0 1.00 . 01 . 97 . 44 . 00
0 . 12 0 0 0 1.00 . 11 . 79 . 91
+ 0 0 . 02 0 0 = 1.00 . 53 . 11
0 0 0 . 11 0 1.00 . 81
[ 0 0 0 0 . 07 ] [ 1.00]
nearly reproduces the correlation matrix R. Thus, on a purely descriptive basis, we
would judge a two-factor model with the factor loadings displayed in the Table as
providing a good fit to the data. The communalities (.98, .88, .98, .89, .93) indicate
that the two factors account for a large percentage of the sample variance of each
variable