Академический Документы
Профессиональный Документы
Культура Документы
YoLt l'taye beefi Gppointed Dedn af Admiss¡ot1s at tlle Mesner School of llealth Care and
Tansorial Tiqcles. your contract stipulates thtit you will receíve a banus of g100,000 eqch year
tllqt the graduati1tl rate exceeds 75aA. Otlly after sigtlilg the contrclct do you fnd thtlt the
success rate for the Last 5 years has averaged only 23.7ok. you decide that the ofily way to
increase this obysmal figure is to impase t¡ghfer adtxissiofis critería, and yotr 111«t wíth the
faculty to drau, up d lkt of the d.sired dtüíbutes of successfuL studet s.TlEy affive.tt three:
(1) lhe eyes of an eagle, (2) tlrc hatl.ls of a watnan, and (3) the soul of a Byzantine usurer. you
det/íse a test battety for appl¡cafits, .,víth five tests i each area, just to be sure you,ve co.¿ered the
areLls well- Uníortunately, the test battery takes )2.6 hours to adnli ister, and you,re still fiof
sure that dll of the tests ín each drea dre tappifi! the right skilk. ls there any way you catl
(1) ruake sure you're fiet$ur¡tlg t¡1ese three areas alxd (2) elífiinate tests thdt are either
redufidant or neasur¡ng somethitlg entirely different?
u.,,ul. $c w,ruldr t br asking rhese qiresrions patlern ol results can be explained by a smaller
IA.I unle\,i ¡h( .rn.wcrs wet( "ye\." Tlre teth- numbcr of underlying constructs (sometimes called
niques wc cover in this chapter to solvc the Dean's latent variables or factors), (l) tcst some hypoth- tTlM(s FA, ot sFA,
dilemma arc principal cornponents analysis eses abour thc data, and (4) reduce the number ol
(PCA) aDd factor analysis (FA).rThey dilfe¡ f¡om vaúables to a more manageable sizc.r In the dark,
techni(lllcs we discussed earlie¡ in one important disfant past, around 0 Bc,a PCA and IA we¡e used for
way: no distinction is rnade berwee¡ independent quite dilferent purposes. Howevet the distinctiorl etllirely.
and dependent va¡iables; all are treated equally and between them has gradually disappeared, and now 2"Some pco¡tle"
neans
are based on one group of subjecls. That is, the goal PCA is used almost exclusively as simply thc lirst we forgot Árho, dfld
ol tlrese tcchniques is to exami[e the s1¡ ucture oÍ the step in FA. We'll keep using both rcrms because --e cd]1't liltd tl1¿
relationship among the variables, not to see how they're still around, and we'll indicate whe¡e one
¡hey relate to other va ables, such as group ment- technique ends and the other begins. )There arc ofher ways
bership or a set ol dependent variables. For rhis So, let's get back to the DeaD's dilemma. After
reasoll, sonle peol)lc have ¡eler¡cd to thesc tech- tlrcsr: lechlliques can
searching the literature for appropriate tests to use,
niqucs as "Lutargcted."2 he comes up with the l5 lisred in the box on page be sell, but we wofi't
To jut[p ahead of the story a bir, our beleaguered Il0, which he administers ro the 200 applicants
Dean will use these two procedures 1(): (1) explore over a l-day period. lThdt's'B¿fore
the relatiorship among rhe variables, (2) see if rhe Compulefi.'
t29
110 REGRESSION AND CORRELAIION
Atkibute Voriohle
Honds
\i..''..
r!.'t.'.'..
\\|ii.:)
\l»¡'¡
'ii$
\l'¡
i '.'
labeled Lr in Figure 15 - 2). we can show this some- the dashed lines is small when compared with that
what more complicated, but accur¿te. pitture in of the solid lines.
Iigure 15-2.
What exactly is meant by 'uniqueness'?6 we can HOW IT'S DONE 6what exactty is rt
best define it in telms of its converse, communal-
fhe Correlation Matrix by any o[ riist
ity. The communality of a variable can be approxi- As we mentioned a bit earliet the first lew steps Howewr, that's ¡t
mated by its multiple corelation, li2, with all oI the in FA. for histodcal reasons, go by the name of question we'd best
other vadables; that is, how much it has in common PCA. We begin with a correlation matrix. on a tech- leave for the
with them and can be predicted by ¿hem. The nical note. we start with a correlation matrix philosophers.
uníqueness for va¡iable I is then simply (l - R,'); mainly because, in our fields, the va ables are each
that portion of variable I that cannot be predicted measured with very dilferent units, so we convert
by (i.e., is unrelated to) the remaining variables. all of them to standard scores. II the variables all
Before we go on, let's complicate the picture just used a simila¡ metric (such as when we factor
a bit more. Figure 15 -2 assumes that factor I plays a analyze items on a test. each using a 0 -to- 7 scale), it
role only for variables I through 5. factor 2 for 6 would be betteL to begin wirh a variance-covariance
through 10, and facfor 3 for 11 through 15. In matrix.
reality, each oI rhe factors influences all of the If life were good to us, we'd prcbably not need to
vadables to some degrce, as in Figure 15-3. We've go any fuÍher than a co[elation matrix; we'd find
added signs of these influences only for the contli- that all of the variables that measure one factor
bution of the first lactor on the other l0 vadables. correlate very strongly with each other and do not
Faclors 2 and J exert a similar influence on the correlate with the measures of the other attribufes
variables, but putting in the lines would have com- (i.e., the picture in Figure 15-2). However, this is
plicated the picture too much. What we hope to find almost never the case. The correlations within a
is that the influence of the factors represented by factor are rarely much above .85, and the measures
. :. ^ .. . ' '...'.
'i ....1
L)2 REGRESSION AND CORRELATION
I ',l
are ahnost always corrcl¿tcd with "unrclated" ones Othcrs, such as SPSS/PC, eive vou irs firsr cousin (on
to soÍle deqrce (morc like Figure l5 l). Thus \"¡e are ils mothe¡'s side), an antiimage cor¡elation ma-
lelt ]ooking ior patterns in a matrix ol In x \ - 1) - 21 trix. This is nothing more than a partial correlalion
uniquc correlations; in our casc, 1 I 5 x 14) : 2, or 105 nratrix with thc siglrs of the oll diagonal clenienrs
(nol cornting lhe 1.00s along thc main diagonal). as Ievcrsed lor somc reasorl that su4lasscth hLloran
shown in Tallle I5 - l. Ncc(llcss to say, rr_ving Lo Íiakc Llnderstanding- In either case, thcy'¡e in¡crltteted in
sense oi thjs jus¡ by e1e is close to impossible. the opposite r'vav as is the correlalio¡ riratrix: a large
flelorc going on to the ncxt step, it's worthí,hile number of high parrial correlations indica¡es \ou
to do a lew "diaenoslic checks" on this corre]ation shouldn't proceed.
r¡atrix. The reason is that compulers are ilrcrcdibly A rclaled di¿gnosiic rest involves Iooking at the
dumb animals. If no undcrll,ing facto¡ial structure communalitics. Because they are the squareLl ,i¿?l¡i
e\i\t, 1.'e\rlrilts ''l lL, ,ott, .rti,,n r.rltir,,,n.irrinq pla correlations, as opposed to pzz¡¡lrl correiations,
of purely random nuinllcrs bctÍ,ccn .30 and +.10 they shor¡ld be above .60 or so, reflecrinlt rhl3 facr
(i.e., prctty closc to 0), with 1.00s on the main di- that the variables are rclalcd to cach other to some
agoual (brcause a vañable is always perfectly corre- degree. You have to be carciul inter?leting rhe
latcd rvith itself), the computer r,/onld still grind commLlnaliLies ir SPSS/PC. The firsl tirlle ir prir-rrs
away merril],, churning oul rcams of paper, Iüll of them out, lhe,v nray (dcpending on other options
numbe¡s and graphs. signilying nolhiDg. The cx 1l,e'll discuss lalerl all Lrc 1.00. Later in the outpur.
treme example (]1 this is an identity matrix, which rherc will bc arothc¡ column of then], rn/ith values
has 1.00s al(»g the nain diagonal and zeros for all ranging [ron] 0.0 ro 1.0; thjs is thc colunln ro look ar.
thc o11 diagonal ierms. So several tests, formal and Arnong the [ornlal sratistica] tests, one ol the
otherwise, have been devek)ped to ensu¡e that so¡re- oldcst is thc Bartlett Test of Sphericity. WithoLrt
1lling is around Io la(tor analyze. going into rhe dctajls ol how it's calcl-rlatcd, i¡ yiclds
Somc ol thc mos¡ Lrselul 'tesls' do not involvc any a chi-square staristic. II its vahte is small, and thc
statistics at all, orher than counting- Tabachnick and associatcd p lcvcl is over.05, thc¡ the cor¡elatioll
'Their book is lilled lidcll (1989)7 recommcnd nothing morc sophisri- matrix doesn'1 di{fcr signiñcantly lrom an idcnri¡y
with utlco¡tnanly catcd rhan an cyeball cireck of the col¡elalior1 ma- matrix and you should slop right thcrc. However,
qaod wi\¿Dtñ dnd t x; if you havc only a lew co¡relations higher lhan Tabachnick and Fidell (1989) statc thai the Barrlerr
should lte ou the shelf .10, savc yoLrr ¡rapcr and sroL) right thcrc. tcst is "notoriously sensitive," especially with large
A sli:lhtly more stringent lest is to look at a ¡tatr';x sarrrple sizes, so cven ifit is statistically signilicant, it
ol lhe partial correlations. This'tcst'is based on docsn't ¡rcao that yott can safely procccd. Conse
the fac¡ that, if the variables do indced correlale quently, Bartlett's rest is a orlc-sided tcst: if ir says
rvith each other because of an undcrlying lactor you should¡¡'l go on to thc P ncipal Components
stmctüre. thcn the correlalion between any two slage, don't; but if it says you can go on, it ain't
variables shonld be.:mall .íter pcúti.Llillg o t the efferÍs ¡eccssarily so.
of the ather ydridbles. Some conputcr p¡ograms, such Anothcr lest is thc Kaiser-Meyer-Olkin Mea-
as BMI)B prinl out the fartial conelation matrix. sure of Sampling Adequacy (usually relerred to
PRINCIPAL COMPONENTS AND FACTOR ANALYS]S
by its nickname, l<MO), \'vhich is based on the No!v, this may seem like a tremendous amount of
squarcd partial corrcl;ltions. Ilt the SPSS/PC com cflort was expended to ge1 absolulely nowhere. lf we
pnlcr packagc. KMO vallre Ic¡¡ each va¡iatrle is began wjth l5 variables a]]d ended up with l5
prinrcd alon8 Ihc mai¡ diago¡al o[ the antiimagc lactors, rvhat have we gailcd? Actlrally, quite a bil.
correl¡lioll malrix, and a sunlmary value is also The ws lor the flrs¡ laclor are chosen so that they
gi\,cn. This allolvs you 1o check 1he overall adequacy cxpress the largest amount of va¡iance in the sarn-
ol ¡hc m¡rrix and also sec rvhich individual va¡iablcs ple. The ws i¡t the secr¡ncl lactor are (lerived to nteet
¡1ay not be pulling thcir lrrll stalislical weighl. Il thc two crilcrla: (I) the second lacto¡ is uncor¡elated
value is i¡ the .60s, Kaiser describes rhe rneasllre as with the lirst, and (2) ir expresses the largest
Tliose values iD lhe.50s are amount of va¡iance lelt over after the iirsl fafior is
ble" and lc¡lvcr oncs arc "rrnacccplable"; you should consldered. The rrs in all the remaining facto¡s arc
procced lvith thc nexl stcp accordingly Similarly, calculated in the same way, with each lactor uncor-
-volr can considcr eliminating valiallles rhal sholv rclated with aud exp)aining less varjance than thc
l, ,,r ., .1' n¡ rdcqua,i. freviou- urr'\. 5o, il J [J.lori,] .tru( lJ rc i. pr('.cnl in
the data, rrost o[ thc variance Ilray be exp]ajned on
Extracting the Factors lhe Lrasis of onll the first ierv lactors.
Assllming that all has gone well in the previor¡s Again returning 1L) our exampie, the Dean hopes
stcps, wc Dol¡/ go on lo extracting the Iactors, a tllat thc first I factors are ¡esponsiblc fo¡ r¡ost of thc
proccdurc only siightly Iess painlul than exlracti¡g variar]ce among the variables and rhat rhc ¡emain
teelh. The purposc of this is to comc r-rp with a series ing 12 factors will be relatively 'wcak' (i.e., he won't
ol linear combinations ol the variables 1o definc lose 1oo nluch information if he ignorcs them). Thc
each laclor. Fo¡ l¿cror l, thjs would look sonethinll actual resl¡lts are given in Table l5-2. For the
like: momcnt, illnore the column headed 'Eigenva]ue'
(we get back to this cryptic word a bil later) and look
I wr2x, - ..- + rrrr,Yr at thc lasL onc, 'Cul1rula¡ive percenr.'Notice that rhe
first laclor accounts for 37.4ok of the va¡iaDce, thc
wberc thc X te¡ms arc the I (in this case, 15) first two for ove¡ 507o, and the first five for almost
variables and the ws a¡e weights. These w tcms 757o of lhe variance of thc original data. So he
have nvo subscripts; thc first shows that lhey go ac¡ually may end up with what he's looking for.
rvlth {actor l, aod the second inclicates with which What we've jus¡ desc¡ibcd is the esscnce of PCA.
varialrle they're associated. The ¡eason is that, il lve what it lries to do. theñ, is explain the variance
have l5 variables, we will end up with 15 fac¡ors among a bunch of va¡ial¡les in terms of uncor¡elated
and therelo¡e 15 equations in the fo¡¡r of the one (the slatistical ferm is orthogonal) underlying fac
above. Fr¡r example, rhe second {actor would look tors or latenl variables. The way ¡t's used now is to
like: lry fo reduce thc rumbcr oI facto¡s as r¡ucl] as
possible so as to get a more parsimonious explana'
F): w)tx\ + ri/22,Y2 + rion of what's going on. In fact, though, PCA is only
134 REGRESSION AND CORRÉLATION
ltTo the extent to which statistician s. I 5 We've simply transformed a number NYSTAGMUS loads strongly on factors I and
stro g emotiofis can of variables into facto$. The only subjective element 2, CARROTS loads olr all I factors to
be arcused ifi statisti- was in selecting the number o{ factors to retain. comparable degrees, and so on. Factorial
cia s (\'rhich is wlly However, if we asked for the factor matrix 10 include complexity makes it more difficult to
\^,e fekr to stdtisti- all of the factors, rather than just those over some interpret the role of the variable. INTEREST
cians as thef, ralher criterion, we could go back and forth between fac- is explained by both factor I and facror 2
tors and vadables without losing any information at and, conversely, the explanation of these
all. It is the next step, Iactor rotation, that really gets factors must take CARROTS into account. Il
the dander up among some (unenlightened) statis- would make life much easie¡ if we could
tical folks. The reason is that we have, literally, an understand the factors on the basis of
inñnite numbe¡ of ways we can rotate the factors. mutually exclusive sets of variables.
which rotation we decrde to use (assuming we don't l. Magnitude of the loadings. This is really a
merely accept the program's default options without consequence oI the second cdterion. If a
question) is totally a matter of choice on the ana- va¡iable loads strongly on one factor, then its
iyst's part. loadings on the other factors will be close to
So, if factor rotation is still somewhat controver- 0. The reason is that the sum oI the squares
sial, why do we do it? Unlike other acts that arouse of the loadings across factors (the va able's
strong passions, we can'a explain it simply on the communal¡ty, you'll remember) remains
basis of the fact that it's fun. To us t¡ue believers, constant when we Iotate; so as some
factor rotation serves some useful functions. The loadings go up, orhe¡s have to go down.
primary one is to help us understand what (if 4. Unipolar factors. If some loadings were
anything) is going on with the factors. positive and others negative, rhen a high
To simplily interpretation of the facton, the fac- score on the factor would indicate more of
tor loading matrix should satisfy lour conditions: some va¡iables, whereas a low score would
l. The variance should be fairly evenly indicate more of other va¡iables. Again, in
distributed across the factors. the inte¡est of interprefive ease, we'd like
2. Each variable should load on only one factor. the factor to be unipolar; rhar is, a higher
l. The factor loadings should be close to l 0 or score on the factor means more of the latent
0.0. va¡iable, and a lowe¡ score simply means less
4. The factors should be unipolar (all the strong oI it. This occurs when all of rhe factor
vari¿bles have ¡he sdme s¡gn). loadings have the same sign.
Let's see how well the {actor loading mat¡ix in From a mathematical viewpoint, nothing is wrong
Table l5-I meets these crite a. with most of the variance being in one factor, or
I. Distribution of variance. If we go back to with factorial complexily, or with loadings in the
Table 15 - 2, we can add up the eigenvalues middle range, or with bipolar factors. Howevet it is
of the first three factois. Their sum, 9.1788, easiest to interpret the iesults of a facto¡ analysis if
shows the amount of variance explained by we can meet these cfite¡ia arrd aim lot sttü¿t tal
them (which is 61.2% of the total va ance síl1lp|ícíty. 'lhis \s what rotating the factors tries ro do.
of l5). Of this amount, the fi¡st Iactor Unfortunately, no one's lound a way to optimize all
accounts for 1.5.6025 + 9.1788). or 61.0%, of these c¡iteria at once. A rotation that spreads the
the second factor Ior 12.0252 + 9.1798) variance equally across the factors may not necessar-
or 22.7ó/", and the third factor for the ily reduce factorial complexity; and one that reduces
remaining 16.9o/o. So, the flIst factor complexity may not produce unipolar facto§. Need-
contains a disproportionate share of the total less to say, this has resulted in a profusion oI rotation
variance explained by the three factors. We techniques. each one designed ro g¡ve priority ro a
can also see this in the {act that all of the diffe¡ent c¡iterion, and all of which yield somewhat
'uwh¿tt fiakes f¡tclot va ables load strongly on this factor (Table different results.r6 The one that's used most is called
l5-3): t2 of the l5 have loadings over.50 varimax, and that's what we'll go with first.
utlique it
the .freld of on factor 1, and only 2 vadables A simple exaÍnple. Let's see how rotating the
stalíslícs is lhtlÍ the (NYSTAGMUS and CARROTS) Ioad higher factors can help meet the four criteria and grant us
l¿chfiiques are not on another factor than they do on lactor l. our wish for simplicity- However, because it's hard to
amed afier people. This situation is extremely common and is draw I -dimensional pictures (l dimension for each
H o1t l¿.,er, after h yí t1 g Iound because consistency tends to occur in factor), we'll start ofI by forcing the PCA to give us
to get !"our ¡onque people across various measures. What factor only two factors. We can then generalize the proce-
!1tout1d lerfis like I often picks up is this "general faoor," dure to three or more Iactors, although we won't be
which only rarely tells us something we able to visualize the results as readily.
binormamin, or didrl't aheady know. By asking lor two facrors, ou¡ facto¡ loading table
oblitltdx, you dliflosl 2. Factorial complexity. whenever a vafiable will have just two columns. Let's plot each va¡iable,
wish they had been loads strongly on two or more factors, we using the loading on Ia«or I as the X coordinate and
gi)en humdrl tlfies. call it factorially complex. In Table l5 - 3, the loading on the second factor as the Fcoordinate.
PRINCIPAL COMPONENTS AND FACTOR ANALYSIS 117
FIGURE 15 - IO
PIot of the factor
loadings ior the
un¡otated and
rotated solutions.
0.0 0.2 0.4 0.ó 0.8 L0 0.0 0.2 0.4 0.ó 0.8 L0
Foctor loodings Foctor loodings
140 REGRESSION AND CORRELATION