© All Rights Reserved

Просмотров: 0

© All Rights Reserved

- Numerical Method for engineers-chapter 19
- 확률통계답안지
- INTEGRATION OF IMPORTANCEPERFORMANCE ANALYSIS AND FUZZY DEMATEL
- Unit 5 ( CORRELATION AND REGRESSION )
- Research
- Credit Expansion and Financial Instability: Evidence from Stock Prices
- Evaluating the Uncertainty of Polynomial Regression Models Using Excel
- Writing an Honours Thesis - Joe Wolfe - UNSW
- Analytical Delay Models for Signalized Intersections Akgungor
- summary of stats
- Do Clicks Count to Increase Student Achievement? Learner Analytics on a Large-Enrollment Hybrid Course - Sponsored furniture provided by Steelcase, Gold Partner (176641159)
- ECO 231 Empirical Project Mets Attendance
- 30 Correlation Main DRAFT
- Assignment 2
- Chapter_01_Regression_5E(1).pdf
- Stats Ch 6 Final Review
- Analysis of Different Urban Forms in Istanbul
- SPSS_DescriptiveInferential
- 2. FORCASTING
- 718849

Вы находитесь на странице: 1из 36

Analyzing Quantitative

Data: Multivariate

Statistics

T usually are complex. Patients’ spirituality,

nurses’ effectiveness, or abrupt elevations of patients’

used to understand the effects of two or more inde-

pendent variables on a dependent variable. The

temperature are phenomena with multiple facets and terms multiple correlation and multiple regres-

multiple determinants. Scientists, in attempting to sion will be used almost interchangeably in refer-

explain or predict phenomena, have recognized that ence to this technique, consistent with the strong

two-variable studies are often inadequate for these bond between correlation and regression. To com-

purposes. The classic approach to data analysis and prehend the nature of this bond, we ﬁrst explain

research design, which involved studying the effect of simple (that is, bivariate) regression.

a single independent variable on a single dependent

variable, is being replaced by more sophisticated

SIMPLE LINEAR

multivariate* procedures. Regardless of whether a

REGRESSION

study is designed for multivariate analysis, the fact

remains that the determinants of most nursing out- Regression analysis is used to make predictions

comes are fundamentally multivariate. about phenomena. In simple (bivariate) regression,

Unlike the statistical methods reviewed in one independent variable (X) is used to predict a

Chapters 19 and 20, multivariate statistics are com- dependent variable (Y). For instance, we could use

putationally formidable. Our purpose is to provide simple regression to predict weight from height,

a general understanding of how, when, and why nursing school achievement from Scholastic

multivariate statistics are used, without working Assessment Test (SAT) scores, or stress from noise

out computations. Those needing more comprehen- levels. An important feature of regression is that the

sive coverage should consult the references at the higher the correlation between two variables,

end of this chapter. the more accurate the prediction. If the correlation

The mostly widely used multivariate proce- between diastolic and systolic blood pressure were

dure (and therefore the one we describe in greatest perfect (i.e., if r 1.00), we would need to mea-

sure only one to know the value of the other.

*We

Because most variables are not perfectly correlated,

use the term multivariate in this chapter to refer to analy-

ses with at least three variables. Some statisticians reserve the predictions made through regression analysis usu-

term for problems involving more than one dependent variable. ally are imperfect.

512 ■ PART 5 The Analysis of Research Data

The basic linear regression equation is The formulas for these two components are as

follows:

Y¿ a bX

xy

where Y prediced value of variable Y b a Y bX

a intercept constant x2

b regression coefﬁcient (slope of the where b regression coefﬁcient

line) a intercept constant

X actual value of variable X Y mean of variable Y

Regression analysis solves for a and b, so for X mean of variable X

any value of X, a prediction about Y can be made. x deviation scores from X

You may remember from high school algebra that y deviation scores from Y

the preceding equation is the algebraic equation for

a straight line. Linear regression is used to deter- Calculations for the current example are pre-

mine a straight-line ﬁt to the data that minimizes sented in columns 3 through 7 of Table 21-1. As

deviations from the line. shown at the bottom of the table, the solution to the

As an illustration, consider the data for ﬁve regression equation is Y 1.5 .9X. Now sup-

subjects on two variables (X and Y) shown in pose for the moment that the X values in column 1

columns 1 and 2 of Table 21-1. These two vari- are the only data we have, and we want to predict

ables are strongly related (r .90). We use the ﬁve values for Y. For the ﬁrst subject, X 1; we would

pairs of X and Y values to solve for a and b in a predict that Y 1.5 (.9)(1), or 2.4. Column 8

regression equation, which will allow us to predict shows Y values for each X. These numbers show

Y values for a new group of people about whom we that Y does not exactly equal the actual values

will have information on variable X only. obtained for Y (column 2). Most errors of

(1) (2) (3) (4) (5) (6) (7) (8) (9) (10)

X Y 2 y y2 y Y e e2

3 6 2 4 0 0 0 4.2 1.8 3.24

5 4 0 0 2 4 0 6.0 2.0 4.00

7 8 2 4 2 4 4 7.8 .2 .04

9 10 4 16 4 16 16 9.6 .4 .16

X 5.0 Y 6.0

r .90

xy 36

b .9

x 2

40

a Y bX 6.0 (.9)(5.0) 1.5

Y¿ a bX 1.5 .9X

CHAPTER 21 Analyzing Quantitative Data: Multivariate Statistics ■ 513

prediction (e) are small, as shown in column 9. the X axis, we must go up 3.6 units (.9 4) on the

Errors of prediction occur because the correlation Y axis. The line, then, embodies the regression

between X and Y is not perfect. Only when r 1.00 equation. To predict a value for Y, we would go to

or 1.00 does Y Y. The regression equation the point on the X axis for an obtained X value, go

solves for a and b in a way that minimizes such up to vertically to the point on the regression line

errors. More precisely, the solution minimizes the directly above the X score, and then read the pre-

sums of squares of prediction errors, which is why dicted Y value horizontally on the Y axis. For

standard regression analysis is said to use a least- example, for an X value of 5, we would predict a Y

squares criterion. Indeed, standard regression is of 6, indicated by the star.

sometimes referred to as ordinary least squares, The correlation coefﬁcient, it may be recalled,

or OLS, regression. In column 10 of Table 21-1, is an index of the degree to which variables are

the error terms—referred to as the residuals—have related. From a statistical point of view, relation-

been squared and summed to yield a value of 7.60. ships are concerned with how variation in one vari-

Any values of a and b other than 1.5 and .9 would able is associated with variation in another. The

have yielded a larger sum of squared residuals. square of r (r2) tells us the proportion of variance

Figure 21-1 displays the solution to this in Y that can be accounted for by X. In our exam-

regression analysis graphically. The actual X and Y ple, r .90, so r2 .81. This means that 81% of

values are plotted on the graph with circles. The the variability in Y values can be understood in

line running through these points represents the terms of variability in X values. The remaining

regression solution. The intercept (a) is the point at 19% is variability resulting from other sources.

which the line crosses the Y axis, which in this case In the regression problem, the sum of the

is 1.5. The slope (b) is the angle of the line. With squared residuals (es) is 7.6. Column 6 of Table

b .90, the line slopes so that for every 4 units on 21-1 also tells us that the sum of the squared

linear regression.

514 ■ PART 5 The Analysis of Research Data

deviations of Y values from the mean of Y (y2) is applicants would have made successful students. It

40, which represents variability in Y. To demon- may be possible, by using additional information,

strate that the residuals are 19% of the unexplained to make more accurate predictions through multi-

variability in Y, we compute the following ratio: ple regression. The basic multiple regression equa-

tion is

7.6

.19 1 r 2

40 Y¿ a b1X1 b2X2 p bkXk

Although the computations are not shown, the where Y predicted value for variable Y

sum of the squared deviations of the Y values from a intercept constant

the mean of Y (y2) equals 32.4. As a proportion k number of predictor (independent)

of the variability in the actual Y values, this would variables

equal b1 to bk regression coefﬁcients for the

k variables

32.4

.81 r 2 X1 to Xk scores or values on the k independent

40 variables

These calculations reinforce a point made ear-

When the number of predictor variables exceeds

lier: The stronger the correlation, the better the pre-

two, computations to solve this equation are very

diction; the stronger the correlation, the greater the

complex and so we restrict our discussion to hypo-

percentage of variance explained.

thetical rather than worked-out examples.

In our example of predicting graduate nursing

M U LT I P L E L I N E A R students’ GPAs, suppose we hypothesized that

REGRESSION information on undergraduate GPA (GPA-U) and

scores on the quantitative portion of the GRE

Because the correlation between two variables is (GRE-Q) would improve our ability to predict

rarely perfect, researchers often try to improve pre- graduate GPA. The resulting equation might be

dictions of Y by including more than one indepen-

dent (predictor) variable. Multiple regression is Y .4 .05(GPA-U)

used for this purpose. .003(GRE-Q) .002(GRE-V)

For instance, suppose an applicant had a GRE-V

Basic Concepts for Multiple score of 600, a GRE-Q score of 550, and a GPA-U of

Regression 3.2. The predicted graduate GPA would be

Suppose we wanted to predict graduate nursing

Y .4 (.05)(3.2)

students’ grade-point averages (GPA). Only a lim-

.003(550) .002(600) 3.41

ited number of students can be accepted, so we

want to select applicants with the greatest chance We can assess the degree to which adding two

of success. Suppose we had previously found that independent variables improves our ability to pre-

students with high scores on the verbal portion of dict graduate school performance through the mul-

the Graduate Record Examination (GRE) tended to tiple correlation coefﬁcient. In bivariate correla-

get better grades than those with lower GRE tion, the index is Pearson’s r. With two or more

scores. The correlation between GRE verbal scores independent variables, the correlation index is the

(GRE-V) and graduate GPAs has been calculated multiple correlation coefﬁcient, or R. Unlike r, R

as .50. With only 25% (.502) of the variance of does not have negative values. R varies only from

graduate GPA accounted for, there will be many .00 to 1.00, showing the strength of the relationship

errors of prediction: Many admitted students will between several independent variables and a

not perform as well as expected, and many rejected dependent variable but not direction. Direction

CHAPTER 21 Analyzing Quantitative Data: Multivariate Statistics ■ 515

cannot be shown because X1 might be positively GRE-V (r .70). All correlations are fairly sub-

related to Y, while X2 might be negatively related. stantial, a fact that helps to explain why R is not

The R statistic, when squared (R2), indicates the much higher than the r between the dependent

proportion of variance in Y accounted for by the variable and undergraduate grades alone (.71 com-

combined, simultaneous inﬂuence of the indepen- pared with .60). This somewhat puzzling phenom-

dent variables. enon results from redundancy of information

R2 provides a way to evaluate the accuracy of among predictors. When correlations among inde-

a prediction equation. Suppose that with the three pendent variables are high, they tend to add little

predictor variables in the current example, the predictive power to each other. With low correla-

value of R .71. This means that 50% (.712) of the tions among predictor variables, each one can con-

variability in graduate students’ grades can tribute something unique to the prediction of the

be explained by their verbal and quantitative GRE dependent variable.

scores and undergraduate grades. The addition of Figure 21-2 illustrates this concept. Each circle

two predictors doubled the variance accounted for represents total variability in a variable. The circle

by the single independent variable (GRE-V), from on the left (Y) is the dependent variable (graduate

.25 to .50. GPA). The overlap between this circle and the other

The multiple correlation coefﬁcient cannot be circles represents variability that the variables have

less than the highest bivariate correlation between in common. If the overlap were complete—if the

an independent variable and the dependent vari- entire Y circle were covered by the other circles—

able. Table 21-2 presents a correlation matrix with R2 would be 1.00. As it is, only half the circle is

the correlation coefﬁcients for all pairs of variables covered, because R2 .50. The hatched area desig-

in this example. The independent variable most nates the independent contribution of undergraduate

strongly correlated with graduate performance is GPA in explaining graduate grades. This contribu-

(GPA-U), r .60. The value of R could not have tion amounts to 36% of Y’s variance (.602). The

been less than .60. remaining two independent variables do not con-

Another important point is that R can be more tribute as much as we would expect by considering

readily increased when the independent variables their bivariate correlation with graduate GPA (r for

have relatively low correlations among them- GRE-V .50; r for GRE-Q .55). In fact, their

selves. In the current case, the lowest correlation combined additional contribution is only 14%

coefficient is between GRE-Q and GPA-U (.50 .36 .14), designated by the dotted area on

(r .40), and the highest is between GRE-Q and the ﬁgure. The contribution is small because the two

GPA-GRAD 1.00

GPA-U .60 1.00

GRE-Q .55 .40 1.00

GRE-V .50 .50 .70 1.00

GPA, grade point average; GRE, Graduate Record Examination; GPA-GRAD, graduate GPA; GPA-

U, undergraduate GPA; GRE-Q, GRE quantitative score; GRE-V, GRE verbal score.

516 ■ PART 5 The Analysis of Research Data

Undergraduate

GPA

X1

Graduate GRE

Y X3

GPA verbal

X2

GRE

quantitative

F I G U R E 2 1 . 2 Visual representation of

multiple regression example.

undergraduate grades.

A related point is that, as more independent Multiple regression analysis is not used solely (or

variables are added to the regression equation, even primarily) to develop prediction equations.

increments to R tend to decrease. It is rare to find Researchers almost always ask inferential ques-

predictor variables that correlate well with a cri- tions about relationships in the regression analy-

terion but modestly with one another. sis (e.g., Is the calculated R the result of chance

Redundancy is difficult to avoid as more and fluctuations, or is does it reflect true relationships

more variables are added to the equation. The in the population?) There are several relevant sig-

inclusion of independent variables beyond the nificance tests, each used to address different

first four or five typically does little to improve questions.

the proportion of variance accounted for or the

accuracy of prediction. Tests of the Overall Equation and R

The dependent variable in multiple regression The most basic test in multiple regression concerns

analysis, as in analysis of variance (ANOVA), the null hypothesis that the population multiple corre-

should be measured on an interval or ratio scale. lation coefﬁcient equals zero. The test for the signiﬁ-

Independent variables, on the other hand, can cance of R is based on principles analogous to those

either be interval- or ratio-level variables or for ANOVA. We stated in Chapter 20 that the general

dichotomous variables. A text such as that by Polit form for calculating an F-ratio is as follows:

(1996) should be consulted for information on

how to use and interpret dichotomous dummy SSbetween dfbetween

F

variables. SSwithin dfwithin

CHAPTER 21 Analyzing Quantitative Data: Multivariate Statistics ■ 517

In the case of multiple regression, the form is predictor increase our ability to predict Y after two

similar: predictors have been used? An F-statistic can be

computed to answer this question.

SSdue to regression/dfregression

F Let us number each independent variable in

SSof residuals/dfresiduals the current example: X1 GPA-U; X2 GRE-Q;

and X3 GRE-V. We can then symbolize various

The basic principle is the same in both cases: correlation coefﬁcients as follows:

Variance resulting from independent variables is

contrasted with variance attributable to other fac- Ry.1 the correlation of Y with GPA-U

tors, or error. To calculate the F-statistic for Ry.12 the correlation of Y with GPA-U and

regression, we can use the following alternative GRE-Q

formula: Ry.123 the correlation of Y with all three

predictors

R2 k

F The values of these Rs are as follows:

(1 R )(N k 1)

2

where k number of predictor variables Ry.12 .71 R2y.12 .50

N total sample size Ry.123 .71 R2y.123 .50

In the example of predicting graduate GPAs, These ﬁgures indicate that GRE-V scores

suppose a multiple correlation coefﬁcient of .71 made no independent contribution to the multiple

(R2 .50) has been calculated for a sample of 100 correlation coefﬁcient. The value of Ry.12 is identi-

graduate students. The value of the F-statistic is cal to the value of Ry.123. Figure 21-2 illustrates

.50 3 this point: If the circle for GRE-V were complete-

F 32.05 ly removed, the area of the Y circle covered by X1

.50 96

and X2 would remain the same.

The tabled value of F with 3 and 96 degrees of We cannot tell at a glance, however, whether

freedom for a signiﬁcance level of .01 is about 4.0; adding X2 to X1 signiﬁcantly increased the prediction

thus, the probability that R .71 resulted from of Y. What we want to know is whether X2 would

chance ﬂuctuations is considerably less than .01. improve predictions in the population, or if its added

Example of multiple regression: predictive power in this sample resulted from chance.

Badr (2001) examined predictors of devel- The F-statistic formula to test the signiﬁcance of

opment among low-birth-weight infants of Latino variables added to the regression equation is

background. The two dependent variables were the (R2Y.12 pk R2Y.12 pk )(k1 k2)

F

1 2

motor and mental development subscales scores of

(1 R2Y.12 pk )(N k1 1)

the Bayley Scales of Infant Development. There 1

were 11 independent variables measuring primarily where R2y.12...k1 squared multiple correlation

maternal factors such as maternal confidence, coefﬁcient for Y correlated

maternal depression, and family income. The value with k1 predictor variables

of R2 was .40 for motor development and .60 for k1 the number of predictors for

mental development scores, both p .01. the larger of the two sets of

predictors

Tests for Adding Predictors R2y.12...k squared R for Y correlated

2

Another question researchers may want to answer with k2 predictor variables

is: Does adding Xk to the regression equation k2 the number of predictors for

signiﬁcantly improve the prediction of Y over that the smaller of the two sets of

achieved with Xk1? For example, does a third predictors

518 ■ PART 5 The Analysis of Research Data

In the current example, the F-statistic for test- in the regression equation, it means that the vari-

ing whether adding GRE-Q scores signiﬁcantly able associated with the coefﬁcient contributes sig-

improves our prediction of Y is niﬁcantly to the regression, even after extraneous

variables are taken into account.

(.50 .36) 1

F 27.16

.50 97

Consulting Table A-2 in Appendix A, we ﬁnd Strategies for Handling Predictors in

that with df 1 and 97 and a signiﬁcance level of Multiple Regression

.01, the critical value is about 6.90. Therefore, adding There are various strategies for entering predictor

GRE-Q to the regression equation signiﬁcantly variables into regression equations. The three most

improved the accuracy of predicting graduate GPA, common are simultaneous, hierarchical, and step-

beyond the .01 level. wise regressions.

Example of adding predictors in multiple

regression: Simultaneous Multiple Regression

McCarter-Spaulding and Kearney (2001) studied The most basic strategy, simultaneous multiple

mothers’ perceptions of insufﬁcient milk. They ﬁrst regression, enters all predictor variables into the

examined demographic predictors of these percep- regression equation at the same time. A single

tions (age, education, and parity), which yielded a regression equation is developed, and statistical

nonsigniﬁcant R2 of .02. They then added a fourth tests indicate the signiﬁcance of R and of individ-

predictor, scores on a parenting self-efﬁcacy scale. ual regression coefﬁcients. The study about pre-

R2 increased to .25; the .23 increment was statisti- dicting infant development cited earlier (Badr,

cally signiﬁcant. 2001) used a simultaneous approach: All 11 inde-

pendent variables were put into the regression

Tests of the Regression Coefﬁcients equation at the same time. This strategy is most

Researchers often want to know the significance appropriate when there is no basis for considering

of individual predictors in the regression equa- any particular predictor as causally prior to anoth-

tion, which can be done by calculating t statistics er, and when the predictors are of comparable

that test the unique contribution of each indepen- importance to the research problem.

dent variable. A significant t indicates that the

regression coefficient (b) is significantly different Hierarchical Multiple Regression

from zero. Many researchers use hierarchical multiple

Regression coefﬁcients in multiple regression regression, which involves entering predictors

must be interpreted differently than coefﬁcients in into the equation in a series of steps. Researchers

bivariate regression. In simple regression, the value control the order of entry, with the order typically

of b indicates the amount of change in predicted based on logical or theoretical considerations. For

values of Y, for a speciﬁed rate of change in X. In example, some predictors may be thought of as

multiple regression, the coefﬁcients represent the causally or temporally prior to others, in which

number of units the dependent variable is predicted case they could be entered in an early step.

to change for each unit change in a given indepen- Another common reason for using hierarchical

dent variable when the effects of other predictors regression is to examine the effect of a key inde-

are held constant. This “holding constant” of other pendent variable after first removing the effect of

variables means that they are statistically con- extraneous variables. The study by McCarter-

trolled, a very important feature that can enhance a Spaulding and Kearney (2001) exemplifies this

study’s internal validity. If a regression coefﬁcient strategy. Demographic characteristics were

is signiﬁcant and extraneous variables are included entered in step 1, and then parenting self-efficacy

CHAPTER 21 Analyzing Quantitative Data: Multivariate Statistics ■ 519

X3

X1

Y Y Y

X2

FIGURE 21.3 Visual representation of stepwise multiple regression analysis.

was entered in step 2. The findings indicated that, cedure continues until no additional predictor sig-

even after taking age, parity, and education into niﬁcantly increases the value of R2.

account, parenting self-efficacy was a significant Figure 21-3 illustrates stepwise multiple regres-

predictor of the mothers’ perceptions about insuf- sion. The ﬁrst variable (X1), has a correlation of .60

ficient milk supply. with Y (r2 .36); no other predictor variable is cor-

With hierarchical regression, researchers related more strongly with the dependent variable.

determine the number of steps as well as the num- The variable X1 accounts for the portion of the vari-

ber of predictors included in each step. When sev- ability of Y represented by the hatched area in step 1

eral variables are added as a block, as in the of the ﬁgure. This hatched area is, in effect, removed

McCarter-Spaulding and Kearney example, the from further consideration, because this portion of Y’s

analysis is a simultaneous regression for those vari- variability is explained or accounted for. Thus, the

ables at that stage. Thus, hierarchical regression variable chosen in step 2 is not necessarily the X vari-

can be considered a controlled sequence of simul- able with the second largest correlation with Y. The

taneous regressions. selected predictor is the one that explains the largest

portion of what remains of Y’s variability after X1 has

Stepwise Multiple Regression been taken into account. Variable X2, in turn, removes

Stepwise multiple regression is a method of a second part of Y so that the independent variable

empirically selecting the combination of indepen- selected in step 3 is the one that accounts for the most

dent variables with the most predictive power. In variability in Y after both X1 and X2 are removed.

stepwise multiple regression, predictors enter the

regression equation in the order that produces the TIP: Stepwise regression is a somewhat con-

greatest increments to R2. The first step selects troversial procedure because variables are

the single best predictor of the dependent variable, entered into the regression equation based solely on sta-

that is, the independent variable with the highest tistical rather than theoretical criteria. Pure stepwise

bivariate correlation with Y. The second variable to regression is best suited to exploratory work, but cross-

enter the equation is the one that produces the validation is essential. If a sample is sufﬁciently large,

largest increase to R2 when used simultaneously we recommend that it be split in half to determine if

with the variable selected in the ﬁrst step. The pro- similar equations result with both subsets of data.

520 ■ PART 5 The Analysis of Research Data

Tarkka, Paunonen, and Laippala (2000) This is because whatever variance the independent

explored factors associated with ﬁrst-time mothers’ variables have in common is attributed to the ﬁrst

ability to cope with their child care responsibilities, variable entered in the analysis.

using stepwise regression with seven predictors Another approach to assessing the relative

(perceived maternal competence, maternal attach- importance of the predictors is to compare regres-

ment to the child, self-concept, relationship with her sion coefﬁcients. Earlier, the regression equation

spouse, breastfeeding status, decision-making sup- was given as

port, and activity level of the child). All were sig-

niﬁcant predictors, and the cumulative R2 was .55. Y¿ a b1X1 b2X2 p bkXk

Perceived competence entered the equation ﬁrst. where b1 to bk regression coefﬁcients

because they are in the units of original scores, which

Scientists want not only to predict phenomena, differ from one X to another. X1 might be in milliliters,

but to explain them. Predictions can be made in X2 in degrees Fahrenheit, and so forth. The use of

the absence of understanding. For instance, in our standard scores (or z scores) eliminates this problem

graduate school example, we could predict per- by transforming all variables to scores with a mean of

formance moderately well without understanding 0.0 and a standard deviation (SD) of 1.00. The for-

the factors contributing to students’ success. For mula for converting raw scores to standard scores* is

practical applications, it may be sufficient to

make accurate predictions, but scientists typically XX

zX

want to understand phenomena and contribute to SDX

knowledge. In standard score form, the regression equation is

In multiple regression, one approach to under-

zY¿ 1zX 2zX p kzX

standing a phenomenon is to explore the relative 1 2 k

importance of independent variables. Unfortunately, where zY¿ predicted value of the standard score

the determination of the relative contributions of for Y

independent variables in predicting a dependent 1 to k standardized regression weights for k

variable is one of the thorniest issues in regression independent variables

analysis. When independent variables are correlated, zX1 to zXk standard scores for the k predictors

as they usually are, there is no ideal way to

With all the s (referred to as beta [ ]

disentangle the effects of variables in the equation.

weights) in the same measurement units, can their

It may appear that the solution is to compare

relative size shed light on how much importance to

the contributions of the Xs to R2. In our graduate

attach to predictor variables? Many researchers

school example, GPA-U accounted for 36% of Y’s

have interpreted beta weights in this fashion, but

variance; GRE-Q explained an additional 14%.

there are problems in doing so. These regression

Should we conclude that undergraduate grades are

coefﬁcients will be the same no matter what the

more than twice as important as GRE-Q scores in

order of entry of the variables. The difﬁculty, how-

explaining graduate school grades? This conclu-

ever, is that regression weights are unstable. The

sion would be inaccurate because the order of entry

values of tend to ﬂuctuate from sample to sam-

of variables in a regression equation affects their

ple. Moreover, when a variable is added to or

apparent contribution. If these two predictor vari-

ables were entered in reverse order (i.e., GRE-Q

ﬁrst), R2 would remain unchanged at .50; however, *Consult a standard statistical textbook for a complete discus-

GRE-Q’s contribution would be .30 (.552), and sion of standard scores.

CHAPTER 21 Analyzing Quantitative Data: Multivariate Statistics ■ 521

subtracted from the regression equation, the beta where N estimated number of subjects needed

weights change. Because there is nothing absolute L tabled value for the desired ␣ and

about the values of the regression coefﬁcients, it is power

difﬁcult to attach theoretical importance to them. k number of predictors

Another method of disentangling relationships

estimated effect size

is through causal modeling, which represents an

important tool in testing theoretical expectations Table 21-3 presents values for L when ␣ .05 and

about relationships. Causal modeling is described power is .50 to .95, in situations in which there are

later in this chapter. up to 20 predictor variables.

As an example, suppose we were planning a

Sample Size for Multiple Regression study to predict functional ability in nursing home

residents using ﬁve predictor variables. We estimate

Small samples are especially problematic in multi- a moderate effect size (R2 .13) and want to achieve

ple regression and other multivariate procedures. a power of .80 and ␣ .05. With R2 .13, the esti-

Inadequate sample size can lead to Type II errors, mated population effect size (

) is .149 (.13 .87).

and can also yield erratic and misleading regres- From Table 21-3, we find that the value of

sion coefﬁcients. L 12.83. Thus:

There are two approaches to estimating sample

size needs. One concerns the ratio of predictor vari- 12.83

N 5 1 92.1

ables to total number of cases. Tabachnick and .149

Fidell (2000) indicate that the barest minimum is to

include ﬁve times as many cases as there are pre- Thus, a sample of about 92 nursing home res-

dictors in the regression equation. They recom- idents is needed to detect a population R2 of .13

mend, however, a ratio of 20 to 1 for simultaneous with ﬁve predictors, with a 5% chance of a Type I

and hierarchical regression and a ratio of 40 to 1 for error and a 20% chance of a Type II error.

stepwise. More cases are needed for stepwise TIP: If you are testing hypotheses and ana-

regression because this procedure capitalizes on lyzing data by computer, you should deﬁ-

the idiosyncrasies of a speciﬁc data set. nitely consider using multivariate procedures in

A better way to estimate sample size needs is lieu of bivariate ones because the commands to the

to perform a power analysis. The number of sub- computer are not appreciably more complicated.

jects needed to reject the null hypothesis that R Researchers can come to the wrong conclusions

equals zero is estimated as a function of effect size, about their hypotheses (i.e., commit Type I errors)

number of predictors, desired power, and the sig- when extraneous variables are not controlled, and

niﬁcance criterion. For multiple regression, the multivariate procedures offer a way to control

estimated population effect size is: them.

R2

1 R2 A N A LY S I S O F

C O VA R I A N C E

Researchers must either predict the value of R2

on the basis of earlier research, or use the conven- Analysis of covariance (ANCOVA) has much in

tion that the effect size will be small (R2 .02), common with multiple regression, but it also has fea-

moderate (R2 .13), or large (R2 .30). Next, the tures of ANOVA. Like ANOVA, ANCOVA is used to

following formula is applied: compare the means of two or more groups, and the

central question for both is the same: Are mean group

L differences likely to be real, or do they reﬂect chance

N k1

ﬂuctuations? Like multiple regression, however,

522 ■ PART 5 The Analysis of Research Data

TABLE 21.3 Power Analysis Table for Multiple Regression: Values of L for .05

Power (1 )

2 4.96 7.70 9.84 10.92 12.65 15.44

3 5.76 8.79 10.90 12.30 14.17 17.17

4 6.42 9.68 11.94 13.42 15.41 18.57

5 6.99 10.45 12.83 14.39 16.47 19.78

6 7.50 11.14 13.62 15.26 17.42 20.86

7 7.97 11.77 14.35 16.04 18.28 21.84

8 8.41 12.35 15.02 16.77 19.08 22.74

9 8.81 12.89 15.65 17.45 19.83 23.59

10 9.19 13.40 16.24 18.09 20.53 24.39

15 10.86 15.63 18.81 20.87 23.58 27.84

20 12.26 17.50 20.96 23.20 26.13 30.72

extraneous variables. This assumption is often violated, however.

Random assignment should be done whenever pos-

Uses of Analysis of Covariance sible, but when randomization is not feasible

ANCOVA can make an important contribution to

ANCOVA is especially useful in certain situations. the internal validity of a study.

For example, if a nonequivalent control group

design is used, researchers must consider whether

Analysis of Covariance Procedures

obtained results are influenced by preexisting

group differences. When experimental control Suppose we were testing the effectiveness of

through randomization is lacking, ANCOVA offers biofeedback therapy on patients’ anxiety. An exper-

post hoc statistical control. Even in true experi- imental group in one hospital is exposed to the

ments, ANCOVA can result in more precise esti- treatment, and a comparison group in another hos-

mates of group differences because, even with pital is not. Subjects’ anxiety levels are measured

randomization, there are typically slight differ- both before and after the intervention and so pretest

ences between groups. ANCOVA adjusts for initial anxiety scores can be statistically controlled

differences so that the results more precisely reﬂect through ANCOVA. In such a situation, the depen-

the effect of an intervention. dent variable is the posttest anxiety scores, the

Strictly speaking, ANCOVA should not be independent variable is experimental/comparison

used with preexisting groups because randomiza- group status, and the covariate is pretest anxiety

CHAPTER 21 Analyzing Quantitative Data: Multivariate Statistics ■ 523

Covariate

X*

variable variable variable

Y Y X

Step 1 Step 2

FIGURE 21.4 Visual respresentation of analysis of covariance.

scores. Covariates can either be continuous vari- of weight change than simple ANOVA. Some

ables (e.g., anxiety test scores) or dichotomous hypothetical data for such a study are displayed in

variables (male/female); the independent variable Table 21-4.

is always a nominal-level variable. Two aspects of the weight values in this table

ANCOVA tests the signiﬁcance of differences are discernible. First, despite random assignment to

between group means after ﬁrst adjusting scores on treatment groups, initial group means are different.

the dependent variable to remove the effect of the Subjects in diet B differ from those in diet C by an

covariate. The adjustment uses regression proce- average of 10 pounds (175 versus 185 pounds).

dures. In essence, the ﬁrst step in ANCOVA is the This difference is not signiﬁcant*; it reﬂects chance

same as the ﬁrst step in hierarchical multiple sampling fluctuations. Second, post-treatment

regression. Variability in the dependent measure means are also different by a maximum of only 10

that can be explained by the covariate is removed pounds (160 to 170). However, the mean number of

from further consideration. ANOVA is performed pounds lost ranged from 10 pounds for diets A and

on what remains of Y’s variability to see whether, B to 25 pounds for diet C.

once the covariate is controlled, signiﬁcant differ- A summary table for ordinary ANOVA is pre-

ences between group means exist. Figure 21-4 sented in part A of Table 21-5. According to the

illustrates this two-step process: Y is patients’ results, group differences in post-treatment means

posttest anxiety scores, the covariate X* is pretest result in a nonsignificant F-value. Based on

anxiety scores, and the independent variable X ANOVA, we would conclude that all three diets

designates whether patients were exposed to the had equal effects on weight loss.

biofeedback therapy. Part B of Table 21-5 presents the summary

Let us consider another example to explore table for ANCOVA. The ﬁrst step of the analysis

further aspects of ANCOVA. Suppose we were breaks total variability in post-treatment weights

testing the effectiveness of three weight-loss diets, into two components: (1) variability explained by

and we randomly assign 30 subjects to 1 of 3

groups. An ANCOVA, using pretreatment weight *Calculations demonstrating nonsigniﬁcance are not shown.

as the covariate, permits a more sensitive analysis The calculated F .45. With 2 and 27 df, p
.05.

524 ■ PART 5 The Analysis of Research Data

180 170 150 140 210 165

160 150 145 135 150 130

215 200 210 190 185 170

155 145 185 185 205 180

205 195 160 150 190 165

175 165 190 175 160 135

180 170 165 150 165 140

140 135 180 165 180 160

195 185 160 160 230 195

XA* 180.0 YA 170.0 X B* 175.0 YB 165.0 X C* 185.0 YC 160.0

X, independent variable (diet); X*, covariate (pretreatment weight); Y, dependent variable (posttreatment weight).

the covariate, pretreatment weights; and (2) residual if appropriate covariates are used. The increase in

variability. The table shows that the covariate sensitivity results from the fact that the covariate

accounted for a signiﬁcant amount of variance, reduces the error term (within-group variability),

which is not surprising given the strong relationship against which treatment effects are compared. In

between pretreatment and post-treatment weights part A of Table 21-5, it can be seen that the within-

(r .91). This merely indicates that people who group term is extremely large (12,300), masking the

started out especially heavy tended to stay that way, contribution of the experimental treatments.

relative to others in the sample. In the second phase Theoretically, it is possible to use any number of

of the analysis, residual variance is broken down to covariates. It is probably unwise, however, to use

reﬂect between-group and within-group contribu- more than ﬁve or six. For one thing, including more

tions. With 2 and 26 degrees of freedom, the F of covariates is often unnecessary because of the typi-

17.54 is signiﬁcant beyond the .01 level. The con- cally high degree of redundancy beyond the ﬁrst few

clusion is that, after controlling for initial weight, variables. Moreover, it is unwise to add covariates that

there is a signiﬁcant difference in weight loss attrib- do not explain a signiﬁcant amount of variability in a

utable to exposure to different diets. dependent variable. Each covariate uses up a degree

This ﬁctitious example was contrived so that the of freedom, and fewer degrees of freedom means that

ANOVA result of “no difference” would be altered a higher computed F is required for signiﬁcance. For

by adding a covariate. Most actual results are less instance, with 2 and 26 df, an F of 5.53 is required for

dramatic than this example. Nonetheless, ANCOVA signiﬁcance at the .01 level, but with 2 and 23 df (i.e.,

yields a more sensitive statistical test than ANOVA adding three covariates), an F of 5.66 is needed.

CHAPTER 21 Analyzing Quantitative Data: Multivariate Statistics ■ 525

SUM OF MEAN

SOURCE OF VARIATION SQUARES df SQUARE f p

Within groups 12,300.0 27 455.6

TOTAL 12,800.0 29

Step 1 c Residual 2236.9 28 79.9

TOTAL 12,800.0 29

Between groups 1284.8 2 642.4 17.54 .01

Step 2 c Within groups 952.1 26 36.6

TOTAL 2236.9 28

TIP: ANCOVA is almost always prefer- after adjusting for pretreatment weights. It some-

able to ANOVA or t-tests, and is relatively times is useful to examine adjusted means, that is,

easy to run on the computer. Covariates are usual- group means on the dependent variable after adjust-

ly available: At a minimum, you can use key back- ing for (i.e., removing the effect of) covariates.

ground characteristics, such as age, gender, and so Means can be adjusted through a process some-

on. Covariates should be variables that you suspect times referred to as multiple classiﬁcation analysis

are correlated with the dependent variable, and (MCA). Adjusted means allow researchers to deter-

demographic characteristics usually are related to mine net effects (i.e., group differences on the

many other attributes. A pretest measure (i.e., a dependent variable that are net of the effect of the

measure of the dependent variable before the covariate.) A subsequent section of this chapter,

occurrence of the independent variable) is an excel- which presents computer examples of multivariate

lent choice for a covariate. statistics, provides an illustration of adjusted means.

Example of ANCOVA:

Adjusted Means

Badger and Collins-Joyce (2000) studied

As shown in part B of Table 21-5, an ANCOVA depression in relation to functional ability in older

summary table provides information about signiﬁ- adults. The sample was divided into two groups

cance tests. The table indicates that at least one of (depressed or not depressed). After statistically

the three groups had a post-treatment weight that is controlling for scores on a physical health impair-

signiﬁcantly different from the overall grand mean, ment measure, they found that those in the

526 ■ PART 5 The Analysis of Research Data

depressed group had signiﬁcantly lower functional Mathematically, a factor is a linear combina-

ability scores than those in the nondepressed group tion of variables in a data matrix, which contains

(F 7.22, df 1,76, p .01). the scores of N people on k different measures. For

instance, a factor might be deﬁned by the following

equation:

F A C T O R A N A LY S I S

The major purpose of factor analysis is to reduce F b1X1 b2X2 b3X3 p bkXk

a large set of variables into a smaller, more man- where

ageable set. Factor analysis disentangles complex F a factor score

interrelationships among variables and identiﬁes X1 to Xk values on the k original variables

variables that “go together” as uniﬁed concepts. b1 to bk weights

This section deals with a type of factor analysis

known as exploratory factor analysis. Another The development of such equations allows us

type—conﬁrmatory factor analysis—uses more to reduce the Xk scores to one or more factor scores.

complex modeling and estimation procedures and

more sophisticated computer programs, as

Factor Extraction

described later.

Suppose we developed 100 Likert-type items Most factor analyses involve two phases. The ﬁrst

measuring women’s attitudes toward menopause. phase (factor extraction) condenses variables in

Our goal is to compare the attitudes of urban ver- the data matrix into a smaller number of factors.

sus rural women. If we do not combine items to The general goal is to extract clusters of highly

form a scale, we would have to perform 100 sepa- interrelated variables from a correlation matrix.

rate statistical tests (such as chi-square tests) to There are various methods of performing the ﬁrst

compare the two groups of women on the 100 step, each of which uses different criteria for

items. We could form a scale by adding together assigning weights to variables. The most widely

scores from several individual items, but which used factor extraction method is called principal

items should be combined? Would it be reasonable components (or principal factor or principal

to combine all 100 items? Probably not, because axes), but other methods include the image, alpha,

the 100 items are not all tapping exactly the centroid, maximum likelihood, and canonical

same thing. There are various aspects to women’s techniques.

attitude toward menopause. One aspect may relate Factor extraction results in an unrotated fac-

to aging, and another to loss of reproductive tor matrix, which contains coefﬁcients or weights

ability. Other questions may involve sexuality, and for all original variables on each extracted factor.

yet others may concern avoidance of monthly (Because unrotated factor matrixes are difﬁcult to

aggravation. There are, in short, multiple dimen- interpret, we postpone a detailed discussion of fac-

sions to women’s attitudes toward menopause, and tor matrixes until the second factor analysis phase

each dimension should be captured on a separate is described.) In the principal components method,

scale. Women’s attitude on one dimension may be weights for the ﬁrst factor are deﬁned such that the

independent of their attitude on another. The iden- average squared weight is maximized, permitting a

tiﬁcation of dimensions can be made a priori by maximum amount of variance to be extracted by

researchers, but different researchers may read dif- the ﬁrst factor. The second factor, or linear combi-

ferent concepts into the items. Factor analysis nation, is formed so that the highest possible

offers an empirical method of clarifying the under- amount of variance is extracted from what remains

lying dimensionality of a large set of measures. after the ﬁrst factor has been taken into account.

Underlying dimensions thus identiﬁed are called The factors thus represent independent sources of

factors. variation in the data matrix.

CHAPTER 21 Analyzing Quantitative Data: Multivariate Statistics ■ 527

FACTOR EIGENVALUE VARIANCE EXPLAINED OF VARIANCE EXPLAINED

2 8.57 23.3 52.5

3 6.91 15.6 68.1

4 2.02 8.4 76.5

5 1.09 6.2 82.7

6 .98 5.8 88.5

7 .80 4.5 93.0

8 .62 3.1 96.1

9 .47 2.2 98.3

10 .25 1.7 100.0

Factor Rotation

ther meaningful variance left, and so a criterion

must be applied to determine when to stop extrac- The second phase of factor analysis (factor rota-

tion and move on to the second phase. There are tion) is performed on factors that have met the

several possible criteria, a fact that makes factor extraction criteria. The concept of rotation can be

analysis a semisubjective process. Several of the best explained graphically. Figure 21-5 shows two

most commonly used criteria can be described by coordinate systems, marked by axes A1 and A2

illustrating information generated in a factor analy- and B1 and B2. The primary axes (A1 and A2)

sis. Table 21-6 presents ﬁctitious values for eigen- represent factors I and II, respectively, as deﬁned

values, percentages of variance accounted for, and before rotation. The points 1 through 6 represent

cumulative percentages of variance accounted for, six variables in this two-dimensional space. The

for 10 factors. Eigenvalues are equal to the sum of weights associated with each variable can be

the squared weights for the factor. Many determined in reference to these axes. For

researchers establish as their cutoff point for factor instance, before rotation, variable 1 has a weight

extraction eigenvalues greater than 1.00. In our of .80 on factor I and .85 on factor II, and variable

example, the ﬁrst ﬁve factors meet this criterion. 6 has a weight of .45 on factor I and .90 on fac-

Another cutoff rule is factors with a minimum of tor II. Unrotated axes account for a maximum

5% explained variance, in which case six factors amount of variance but rarely provide a structure

would qualify in this example. Yet another criterion with conceptual meaning. Interpretability is

is based on a principle of discontinuity: A sharp enhanced by rotating the axes so that clusters of

drop in the percentage of explained variance variables are distinctly associated with a factor. In

indicates the appropriate termination point. In the ﬁgure, B1 and B2 represent rotated factors.

Table 21-6, we might argue that there is consider- The rotation has been done in such a way that vari-

able discontinuity between the third and fourth fac- ables 1, 2, and 3 have large weights on factor I and

tors. The general consensus is that it is better to small weights on factor II, and the reverse is true

extract too many factors than too few. for variables 4, 5, and 6.

528 ■ PART 5 The Analysis of Research Data

factor rotation.

Researchers choose from two classes of rota- cating oblique rotation point out that if the

tion. Figure 21-5 illustrates orthogonal rotation, concepts are correlated, then the analysis should

in which factors are kept at right angles to one reﬂect this fact. In practice, similar conclusions are

another. Orthogonal rotations maintain the inde- often reached by both rotation procedures.

pendence of factors—that is, orthogonal factors are Researchers work with a rotated factor matrix

uncorrelated with one another. Oblique rotations, in interpreting the factor analysis. An example of

on the other hand, permit rotated axes to depart such a matrix is displayed in Table 21-7. To make

from a 90-degree angle. In our ﬁgure, an oblique the discussion less abstract, let us say that the 10

rotation would have put axis B1 between variables variables listed in the table are the ﬁrst 10 of the 100

2 and 3 and axis B2 between variables 5 and 6. items measuring attitudes toward menopause, and

This placement strengthens the clustering of that 3 factors have been extracted and rotated. The

variables around an associated factor, but results in entries under each factor are the weights, called fac-

correlated factors. Some researchers insist that tor loadings. For orthogonally rotated factors, factor

orthogonal rotation leads to greater theoretical clar- loadings can range from 1.00 to 1.00 and can be

ity; others claim that it is unrealistic. Those advo- interpreted like correlation coefficients—they

CHAPTER 21 Analyzing Quantitative Data: Multivariate Statistics ■ 529

2 .02 .61 .18

3 .59 .11 .03

4 .05 .08 .48

5 .21 .79 .02

6 .07 .52 .29

7 .08 .19 .80

8 .68 .12 .01

9 .04 .08 .61

10 .43 .13 .06

express the correlations between individual vari- indicate 5 underlying dimensions. By reducing 100

ables and factors (underlying dimensions). In this variables to 5 new variables, the analysis of differ-

example, variable 1 is fairly highly correlated with ences between urban and rural women using t-tests

factor 1, .75. By examining factor loadings, it is pos- or ANCOVA would be greatly simpliﬁed.

sible to ﬁnd which variables “belong” to a factor. Several methods can be used to form factor

Variables 1, 3, 8, and 10 have fairly sizable loadings scores, one of which is to weight variables accord-

on factor I. (Loadings with an absolute value of ing to their factor loading on each factor. For exam-

about .40 or .30 normally are used as cutoff values.) ple, a factor score for the ﬁrst factor in Table 21-7

The underlying dimensionality of the data can then could be computed for each subject by the follow-

be interpreted. By inspecting items 1, 3, 8, and 10, ing formula:

we would search for a common theme that makes

the variables go together. Perhaps these four ques- F1 .75X1 .02X2 .59X3 .05X4

tions have to do with the link between menopause .21X5 .07X6 .08X7 .68X8 .04X9

and infertility, and perhaps items 2, 5, and 6 (items .43X10 p b100X100

with high loadings on factor II) are related to sexu- where the X values* subjects’ scores on the 100

ality. The naming of factors is a process of identify- items

ing underlying constructs.

Most factor analysis computer programs can

Factor Scores directly compute this type of factor score.

A second method of obtaining factor scores is

When the main purpose of factor analysis is to to select variables to represent the factor and to

delineate the conceptual makeup of a set of mea- combine them with unit weighting. This method

sures, the analysis may end at this point. Frequently,

however, researchers want to develop factor scores *Standardized (z) scores are often used in lieu of raw scores in

to use in subsequent analyses. For instance, the computing factor scores when the means and standard devia-

factor analysis of the 100 menopause items might tions of the variables differ substantially.

530 ■ PART 5 The Analysis of Research Data

would produce the following factor score on the practitioners, whose responses were factor ana-

ﬁrst factor of our example: lyzed. An eigenvalue of 1.00 was used as the

extraction criterion; 33 items with factor loadings

F1 X1 X3 X8 X10 p less than .35 were discarded. The remaining items

The variables are those with high loadings on fac- formed six factors: collegiality; autonomy; interac-

tor I (the “? . . .” indicates that some of the other 90 tions; professional growth; and beneﬁts.

items with high loadings would be included). This pro-

cedure is computationally and conceptually simpler O T H E R L E A S T- S Q U A R E S

than the ﬁrst approach, and yields similar results. M U L T I VA R I A T E

To illustrate more concretely, consider the TECHNIQUES

rotated factor matrix in Table 21-7 and assume that

factor scores are being computed on the basis of In this section, the methods known as discriminant

this 10-item analysis, omitting for the sake of sim- function analysis, canonical correlation, and multi-

plicity the remaining 90 items on the menopause variate analysis of variance are introduced. The

scale. Suppose two respondents had the following introduction is brief, and computations are entirely

scores on the ﬁrst 10 items: omitted because these procedures are very complex.

The intent is to acquaint you with the types of

Subject 1: 7, 1, 1, 5, 1, 6, 7, 6, 3, 2 research situations for which these methods are

Subject 2: 2, 5, 6, 3, 7, 3, 2, 3, 4, 4 appropriate. Advanced statistical texts such as the

Factor I has high loadings on variables 1, 3, 8, ones listed in the references may be consulted for

and 10. Thus, the two subjects’ factor scores on fac- more information.

tor 1 would be

Discriminant Analysis

7 1 6 2 10

2 6 3 4 5 In multiple regression, the dependent variable is

normally a measure on either the interval or ratio

The minus signs reﬂect the negative loadings scale. The regression equation makes predictions

on variables 3 and 10.* The same procedure would about scores that take on a range of values, such as

be performed for all three factors, yielding the fol- heart rate, weight, or scores on a depression scale.

lowing factor scores: Discriminant analysis, in contrast, makes predic-

Subject 1: 10, 3, 9 tions about membership in categories or groups.

Subject 2: 5, 9, 1 The purpose of the analysis is to distinguish groups

from one another on the basis of available indepen-

Factor scores would be computed for all dent variables. For instance, we may wish to pre-

respondents, and these scores could be used in sub- dict membership in such groups as complying

sequent analyses as measures of different dimen- versus noncomplying cancer patients, graduating

sions of attitudes toward menopause. nursing students versus dropouts, or normal preg-

Example of factor analysis: nancies versus those terminating in a miscarriage.

Misener and Cox (2001) developed a scale Discriminant analysis develops a regression

to measure nurse practitioner job satisfaction. They equation—called a discriminant function—for a

administered 77 items to a sample of 342 nurse categorical dependent variable, with independent

variables that are either dichotomous or continu-

ous. Researchers begin with data from subjects

*Researchers forming scale scores with Likert items often reverse

the direction of the scoring on negatively loaded items before whose group membership is known, and develop

forming factor scores. Direction of an item can be reversed by an equation to predict membership when only mea-

subtracting the raw score from the sum of 1 the maximum item sures of the independent variables are available.

value. For example, in a ﬁve-point scale, a score of 2 would be

reversed to 4 [(1 5) 2]. When such reversals have been done, The discriminant function indicates to which group

all raw scores can be added to compute factor scores. each subject will probably belong.

CHAPTER 21 Analyzing Quantitative Data: Multivariate Statistics ■ 531

Discriminant analysis for predicting member- gap between multiple regression and canonical

ship into only two groups (e.g., nursing school correlation is greater than this statement suggests.

dropouts versus graduates) is relatively simple and Like other techniques described thus far,

can be interpreted in much the same way as canonical correlation uses the least-squares prin-

multiple regression. When there are more than two ciple to partition and analyze variance. Basically,

groups, the calculations and interpretations are two linear composites are developed, one of

more complex. With three or more groups (e.g., which is associated with the dependent variables,

very—low-birth-weight, low-birth-weight, and nor- the other of which is for the independent vari-

mal-birth-weight infants), the number of discrimi- ables. The relationship between the two linear

nant functions is either the number of groups minus composites is expressed by the canonical corre-

1 or the number of independent variables, whichev- lation coefficient, RC, which, when squared,

er is smaller. The ﬁrst discriminant function is the indicates the proportion of variance accounted

linear combination of predictors that maximizes the for in the analysis. When there is more than one

ratio of between-group to within-group variance. dimension of covariation in the two sets of vari-

The second function is the linear combination that ables, more than one canonical correlation can be

maximizes this ratio, after the effect of the ﬁrst found.

function is removed. Because independent variables Examples of research using canonical correla-

have different weights on the various functions, it is tion are not common, perhaps because of its com-

possible to develop theoretical interpretations based plexity and perhaps because it is problematic as a

on the knowledge of which predictors are important hypothesis-testing technique.

in discriminating among different groups.

Example of canonical correlation:

As with multiple regression analysis, it is possi-

Lucas, Orshan, and Cook (2000) used

ble to use a stepwise approach to enter predictors into

canonical correlation to explore various compo-

a discriminant analysis equation. Discriminant analy-

nents of Pender’s Health Promotion Model. They

sis also produces an index designating the proportion

studied the role of cognitive-perceptual factors

of variance in the dependent variable accounted for

(e.g., self-esteem, perceived health) and modifying

by predictor variables. The index is Wilks’ lambda

factors (e.g., age, income) in various health-

(), which actually indicates the proportion of vari-

promoting behaviors (e.g., stress management)

ance unaccounted for by predictors, or 1 R2.

among community-dwelling older women. Two

Example of discriminant analysis: dimensions of covariation were identified; the

Aminzadeh and Edwards (2000) examined value of RC for the ﬁrst canonical variate was .74.

predictors of cane use among community-dwelling

older adults. Using a stepwise discriminant analysis,

they determined that age, subjective norms, and atti- Multivariate Analysis of Variance

tudes predicted cane use. The discriminant function

Multivariate analysis of variance (MANOVA) is

accounted for 67% of the variance in cane use

the extension of ANOVA to more than one depen-

( .33), and correctly categorized 91% of the cases.

dent variable. MANOVA is used to test the signif-

icance of differences in group means for two or

more dependent variables, considered simultane-

Canonical Correlation

ously. For instance, if we wanted to examine the

Canonical correlation analyzes the relationship effect of two methods of exercise on both diastolic

between two or more independent variables and and systolic blood pressure, MANOVA would be

two or more dependent variables. Conceptually, appropriate. Researchers often analyze such data

one can think of this technique as an extension of by performing two separate univariate ANOVAs.

multiple regression to more than one dependent Strictly speaking, this practice is incorrect.

variable. Mathematically and interpretatively, the Separate ANOVAs imply that the dependent

532 ■ PART 5 The Analysis of Research Data

another when, in fact, they have been obtained Multivariate Techniques

from the same subjects and are correlated.

MANOVA takes the intercorrelations of dependent We could state an analogy that canonical correla-

variables into account in computing the test statis- tion is to multiple regression what MANOVA is to

tics. However, ANOVA is a more widely under- ANOVA. This analogy, although correct, obscures

stood procedure than MANOVA, and thus its a point that astute readers have perhaps suspected,

results may be more easily communicated to a and that is the close afﬁnity between multiple

broad audience. regression and ANOVA.

MANOVA can be readily extended in ways In fact, ANOVA and multiple regression are

analogous to ANOVA. For example, it is possible virtually identical. Both techniques analyze total

to perform multivariate analysis of covariance variability in a dependent measure and contrast

(MANCOVA), which allows for the control of variability attributable to independent variables

extraneous variables (covariates) when there are with that attributable to random error. Both multi-

two or more dependent variables. MANOVA can ple regression and ANOVA boil down to an F-ratio.

also be used when there are repeated measures of By tradition, experimental data typically are ana-

the dependent variables. lyzed by ANOVA, and correlational data are ana-

lyzed by regression. Nevertheless, any data for

which ANOVA is appropriate can also be analyzed

TIP:One problem with using multivariate

by multiple regression, although the reverse is not

procedures is that the findings are less

true. Multiple regression often is preferable

accessible to statistically unsophisticated readers,

because it is more ﬂexible and provides more infor-

such as practicing nurses. You may thus opt to use

mation, such as a prediction equation and an index

simpler analyses to enhance the utilizability of the

of association, R. (However, most ANOVA pro-

ﬁndings (e.g., three separate ANOVAs rather than a

grams also indicate the value eta-squared, which

MANOVA). If the primary reason for not using

also summarizes the strength of association

appropriate multivariate procedures is for commu-

between variables.)

nication purposes, you should consider running the

analyses both ways. You could present bivariate

results (e.g., from an ANOVA) in the report, but

then note whether a more complex test (e.g. CAUSAL MODELING

MANOVA) changed the conclusions. Causal modeling involves the development and

testing of a hypothesized causal explanation of a

phenomenon, typically with data from nonexperi-

Example of MANOVA: mental studies. In a causal model, researchers posit

Fuller (2000) studied infant behaviors in explanatory linkages among three or more vari-

relation to infant age, distress, and level of pain ables, and then test whether hypothesized pathways

using a three-factor (2 2 4) MANOVA. The from the causes to the effect are consistent with the

dependent variables included several measures of data. We brieﬂy describe some features of two

infants’ behavioral state (e.g., body movements) approaches to causal modeling without discussing

and four cry measures (e.g., pitch, cry energy). analytic procedures.

There were numerous signiﬁcant main effects for

all three factors, as well as interaction effects. For

example, mean differences in least distressed ver-

Path Analysis

sus most distressed infants with regard to “arching

back,” “arm waving,” and “knitted brows” were all Path analysis, which relies on multiple regression,

signiﬁcant at the .01 level. is a method for studying causal patterns among

CHAPTER 21 Analyzing Quantitative Data: Multivariate Statistics ■ 533

variables. Path analysis is not a method for discov- (V1) and illness severity (V2) are exogenous; no

ering causes; rather, it is a method applied to a pre- attempt is made in the model to elucidate what

speciﬁed model formulated on the basis of prior causes different nursing actions or different

knowledge and theory. degrees of illness. An endogenous variable, by

Path analytic reports often display results in a contrast, is one whose variation is determined by

path diagram; we use such a diagram (Figure 21-6) other variables in the model. In our example, self-

to illustrate key concepts. This model postulates that care capacity (V3) and length of hospitalization

the dependent variable, patients’ length of hospitaliza- (V4) are endogenous.

tion (V4), is the result of their capacity for self-care Causal linkages are shown on a path diagram

(V3); this, in turn, is affected by nursing actions (V1) by arrows drawn from presumed causes to pre-

and the severity of their illness (V2). This model is a sumed effects. In our illustration, severity of illness

recursive model, which means that the causal ﬂow is is hypothesized to affect length of hospitalization

unidirectional and without feedback loops. In other both directly (path p42) and indirectly through the

words, it is assumed that variable 2 is a cause of vari- mediating variable self-care capacity (paths p32

able 3, and that variable 3 is not a cause of variable 2. and p43). Correlated exogenous variables are indi-

In path analysis, a distinction is made between cated by curved lines, as shown by the curved line

exogenous and endogenous variables. An exoge- between nursing actions and illness severity.

nous variable is a variable whose determinants lie Ideally, the model would totally explain the

outside the model. In Figure 21-6, nursing actions outcome of interest (i.e., length of hospitalization).

V1

Nursing

actions

P 31

r12

V3

P43 V4

Capacity

Length of

for

hospital stay

P 32 self-care

V2

Severity

of

P 42

illness

P4e4

P3e3

e3 e4

534 ■ PART 5 The Analysis of Research Data

In practice, this almost never happens because effects. Polit (1996) and Pedhazur (1997) provide

there are other determinants of the outcome, an overview of path analytic techniques, and

referred to as residual variables. In Figure 21-6, advanced students should refer to books such as the

there are two boxes labeled e, which denote a com- classic text by Blalock (1972).

posite of all determinants of self-care capacity and

hospital stay that are not in the model. If we can Example of path analysis:

identify and measure additional causal forces and Berger and Walker (2001) tested an explana-

incorporate them into the theory, they should be in tory model of fatigue in women receiving adjuvant

the model. breast cancer chemotherapy. The predictor variables

Path analysis solves for path coefﬁcients, in the model included health and functional status,

which are the weights representing the effect of one type of chemotherapy protocol, health-promoting

variable on another. In Figure 21-6, causal paths lifestyle behaviors, nutritional status, interpersonal

indicate that one variable (e.g., V3) is caused by relations, symptom distress, and initial reaction to

another (e.g., V2), yielding a path labeled p32. In the diagnosis of cancer. In the ﬁnal model, symptom

research reports, path symbols would be replaced distress had the strongest direct path to perceived

by actual path coefﬁcients, which are derived fatigue 48 hours after treatment.

through regression procedures. Path coefﬁcients are

standardized partial regression slopes. For example,

path p32 is equal to 32.1—the beta weight between Linear Structural Relations

variables 2 and 3, holding variable 1 constant. Analysis—LISREL

Because path coefﬁcients are in standard form, they A drawback of path analysis using ordinary least-

indicate the proportion of a standard deviation dif- squares regression is that the method’s validity is

ference in the caused variable that is directly attrib- based on a set of restrictive assumptions, most of

utable to a 1.00 SD difference in the speciﬁed which are virtually impossible to meet. First, it is

causal variable. Thus, the path coefﬁcients give us assumed that variables are measured without error,

indication about the relative importance of various but (as discussed in Chapter 18) most measures

determinants. contain some error. Second, it is assumed that

Path analysis estimates path coefficients residuals (error terms) in the different regression

through the use of structural equations. The equations are uncorrelated. This assumption is sel-

structural equations for Figure 21-6 are as follows: dom tenable because error terms often represent

z1 e1 unmeasured individual differences—differences

z2 e2 that are not entirely random. Third, traditional path

z3 p31z1 p32z2 e3; and analysis assumes that the causal ﬂow is unidirec-

z4 p43z3 p42z2 e4 tional or recursive. In reality, causes and effects are

often reciprocal or iterative.

These equations indicate that z1 and z2 (stan- Linear structural relations analysis, usually

dard scores for variables 1 and 2) are determined by abbreviated as LISREL,* is a more general and

outside variables (es); that z3 depends directly on z1 powerful approach that avoids these problems.

and z2 plus outside variables; and that z4 depends Unlike path analysis, LISREL can accommodate

directly on z2, z3, and outside variables. These struc- measurement errors, correlated residuals, and

tural equations could be solved to yield path coefﬁ-

cients by performing two multiple regressions: by

regressing variable 3 on variables 1 and 2; and by

regressing variable 4 on variables 2 and 3. *LISREL is actually the name of a speciﬁc computer program

for analyzing covariance structures and for performing structur-

Path analysis involves a number of procedures al equations modeling, but the software has come to be almost

for testing causal models and for determining synonymous with the approach.

CHAPTER 21 Analyzing Quantitative Data: Multivariate Statistics ■ 535

nonrecursive models that allow for reciprocal cau- measured variables. The measurement model is

sation. Another attractive feature of LISREL is that essentially a factor analytic model that seeks to

it can be used to analyze causal models involving conﬁrm a hypothesized factor structure. Thus,

latent variables. A latent variable is an unmea- loadings on the factors (the latent variables) pro-

sured variable corresponding to an abstract vide a method for evaluating relationships between

construct. For example, factor analysis yields infor- observed and unobserved variables.

mation about underlying, latent dimensions that are We illustrate with a simpliﬁed example involv-

not measured directly. In LISREL, latent variables ing the relationship between two latent variables: a

are captured by two or more measured variables student’s cognitive ability and academic success. In

(manifest variables) that are indicators of the our example, depicted in Figure 21-7, cognitive

underlying construct. ability is captured by two indicators: scores on the

Path analysis uses least-squares estimation verbal and quantitative portions of the GRE.

with its associated assumptions and restrictions. According to the model, people’s test scores are

LISREL uses a different estimation procedure caused by their cognitive ability (and thus the

known as maximum likelihood estimation. arrows indicating a hypothesized causal path), and

Maximum likelihood estimators are ones that esti- are also affected by error (e1 and e2). Moreover, it is

mate the parameters most likely to have generated hypothesized that the error terms are correlated, as

the observed measurements. indicated by the curved line connecting e1 and e2.

LISREL proceeds in two phases. In the ﬁrst Correlated measurement errors on the GRE might

phase, which corresponds to a conﬁrmatory factor arise as a result of the person’s test anxiety or level

analysis (CFA), a measurement model is tested. of fatigue—factors that would systematically

When there is evidence of an adequate ﬁt of the data depress test scores on both parts of the examination.

to the hypothesized measurement model, the causal The second latent variable is academic success,

model is tested by structural equation modeling. which is captured by undergraduate and graduate

GPAs. The error terms associated with these two

The Measurement Model Phase manifest variables are presumed not to be correlat-

The measurement model stipulates the hypothe- ed but, within the measurement model, the two

sized relationships among latent and manifest vari- latent variables are hypothesized to be correlated.

ables. In path analysis, it is assumed—unrealistical- The hypothesized measurement model then

ly—that underlying constructs on the one hand and would be tested against actual data using LISREL.

the scales used to measure them on the other are The analysis would provide loadings of the

identical. In LISREL, constructs are treated as unob- observed variables on the latent variables, the corre-

served variables measured by two or more fallible lation between the two latent variables, and the cor-

indicators. Thus, a major strength of LISREL is that relation between e1 and e2. The analysis would also

it enables researchers to separate latent variables indicate whether the overall model ﬁt is good, based

from errors, and then to study causal relationships on a goodness-of-ﬁt statistic. If the hypothesized

among latent variables. The process depends on the model is not a good ﬁt to the study data, the mea-

judicious selection of indicators. A faulty selection surement model could be respeciﬁed and retested.

of measured variables to represent a latent variable

leads to questions about whether the theory’s con- The Structural Equations Model Phase

structs are really embedded in the causal model. After an adequate measurement model has been

During the measurement model phase, LIS- found, the second phase of LISREL can proceed. As

REL tests three things: (1) causal relationships in path analysis, researchers specify the theoretical

between measured and latent variables, (2) correla- causal model to be tested. As an example, we could

tions between pairs of latent variables, and (3) use LISREL to test a model in which we

correlations among the errors associated with the hypothesized that the latent variable cognitive ability

536 ■ PART 5 The Analysis of Research Data

Cognitive Academic

ability success

verbal quantitative GPA GPA

scores scores

e1 e2 e3 e4

caused the latent variable academic success. The LISREL has gained considerable popularity

model being tested would look similar to the one in among nurse researchers, but is a highly complex

Figure 21-7, except that there would be a straight procedure. Readers interested in further informa-

line and an arrow (i.e., a path) from cognitive ability tion on LISREL are urged to consult Jöreskog and

to academic success. There would also be an arrow Sörbom (1997), Kelloway (1998), or Bollen and

leading to academic success from a residual term. Bollen (1989).

In this part of the analysis, LISREL yields

information about the hypothesized causal parame- Example of LISREL:

ters—that is, the path coefﬁcients, which are pre- Mahon and Yarcheski (2001) used LISREL

sented as beta weights. The coefﬁcients indicate the to test two models predicting health outcomes and

expected amount of change in the latent endoge- interpersonal outcomes in relation to depression

nous variable that is caused by a change in the latent among early adolescents. In both models,

causal variable. The LISREL program provides depressed mood contributed fairly strongly to

information on the signiﬁcance of individual paths. depressive symptoms, which in turn affected health

The residual terms—the amount of unexplained and interpersonal outcomes.

variance for the latent endogenous variables—can

also be calculated from the LISREL analysis. The

O T H E R M U L T I VA R I A T E

overall ﬁt of the causal model to the research data

S TAT I S T I C A L

can be tested by means of several alternative statis-

PROCEDURES

tics. Two such statistics are the goodness-of-ﬁt

index (GFI) and the adjusted goodness-of-ﬁt The statistical procedures described in this and the

index (AGFI). For both indexes, a value of .90 or previous two chapters cover most of the techniques

greater indicates a good ﬁt of the model to the data. for analyzing quantitative data used by nurse

CHAPTER 21 Analyzing Quantitative Data: Multivariate Statistics ■ 537

computers and new developments in statistical analy- Cowan, Pike, and Budzynski (2001) used

sis have combined to give researchers more options Cox proportional hazards regression to assess the

for analyzing their data than in the past. Although a impact of psychosocial nursing therapy after sud-

full exposition of other sophisticated statistical pro- den cardiac arrest on 2-year survival. The risk of

cedures is beyond the scope of this book, we brieﬂy cardiovascular death was signiﬁcantly reduced by

describe a few advanced techniques and provide ref- the intervention, even after controlling for such fac-

erences for those interested in fuller discussion. tors as depression, prior myocardial infarction, and

decreased heart rate variability.

Survival and Event History Analysis

Logistic Regression

Survival analysis (or life table analysis) is widely

used by epidemiologists and medical researchers Logistic regression (sometimes referred to as logit

when the dependent variable is a time interval analysis) uses maximum likelihood estimation for

between an initial event and a terminal event (e.g., analyzing relationships between multiple indepen-

when the initial event is disease onset or initiation dent variables and a categorical dependent variable.

of treatment, and death is the end event). Survival Logistic regression transforms the probability of an

analysis calculates a survival score, which com- event occurring (e.g., that a woman will practice

pares survival time for one subject with that for breast self-examination) into its odds. The odds of

other subjects. When researchers are interested in an event is the ratio of two probabilities: the prob-

group comparisons—for example, comparing the ability of an event occurring to the probability that

survival function of subjects in an experimental it will not occur. Probabilities that range between

group versus a control group—a statistic can be zero and one in actuality are transformed into con-

computed to test the null hypothesis that the groups tinuous variables that range between zero and inﬁn-

are samples from the same survival distribution. ity. Because this range is still restricted, a further

Survival analysis can be applied to many situ- transformation is performed, namely calculating

ations unrelated to mortality. For example, survival the logarithm of the odds. The range of this new

analysis could be used to analyze such time-related variable (known as the logit) is from minus to plus

phenomena as length of time in labor, length of inﬁnity. Using the logit as the dependent variable, a

time elapsed between release from a psychiatric maximum likelihood procedure estimates the coef-

hospital and reinstitutionalization, and length of ﬁcients of the independent variables.

time between the termination of a ﬁrst pregnancy Logistic regression is based on the assumption

and the onset of a second one. Further information that the underlying relationships among variables is

about survival analysis can be found in Harrell an S-shaped probabilistic function—an assumption

(2001) and Lee (1992). that is usually more tenable than the least-squares

More recently, extensions of survival analysis assumption of linearity and multivariate normality.

have been developed that allow researchers to exam- Logistic regression thus is often more technically

ine determinants of survival-type transitions in a mul- appropriate than discriminant analysis, which also

tivariate framework. In these analyses, independent has categorical dependent variables. In practice,

variables are used to model the risk (or hazard) of logistic regression and discriminant analysis often

experiencing an event at a given point in time, given yield the same results. Logistic regression, however,

that one has not experienced the event before that is better suited to many research questions because it

time. The most common speciﬁcation of the hazard is models the probability of an outcome rather than

known as the Cox proportional hazards model. predicting group membership. Also, logistic regres-

Further information may be found in Therneau and sion enables the researcher to generate odds ratios

Grambsh (2000) and Hougaard (2000). that can be meaningfully interpreted and graphically

538 ■ PART 5 The Analysis of Research Data

displayed. An odds ratio (OR) is the ratio of one prior pregnancies. Figure 21-8 presents part of the

odds to another odds, and thus is an index of relative Statistical Package for the Social Sciences (SPSS)

risk—that is, the risk of an event occurring given one printout for a stepwise multiple regression in which

condition, versus the risk of it occurring given a dif- infant birth weight is the dependent variable and

ferent condition. Logistic regression was rarely used maternal age, smoking status, and number of prior

20 years ago because of the lack of convenient com- pregnancies are predictor variables. We will

puter programs, but that is no longer the case. explain some of the most noteworthy aspects of

this printout.

Example of logistic regression:

The top panel of Figure 21-8 shows that the

Nyamathi, Leake, Longshore, and Gelberg

ﬁrst variable to enter the regression equation is

(2001) examined the congruity between homeless

mother’s age. The second panel (Model Summary)

women’s self-reports of cocaine use and hair

indicates that R .594–the same value shown in

assays for the presence of cocaine. They used

Figure 20-6 as the bivariate correlation. The value

logistic regression analysis to predict underreport-

of R2 is .353 (.5942), which represents the propor-

ing of cocaine use. Being Latina, younger, and liv-

tion of variance in birth weight accounted for by

ing primarily in shelters were signiﬁcant predictors

mother’s age.* Maternal age was the ﬁrst of the

of underreporting. For example, the odds of under-

reporting were twice as high among women whose three predictor variables stepped into the equation

usual living place in the prior month was shelters because it had the highest bivariate correlation with

than among other women (OR 2.01, p .05). birth weight. (The r between birth weight and num-

ber of pregnancies is .32, and the r between birth

weight and smoking status is .24.)

THE COMPUTER AND The next panel (ANOVA) shows the calcula-

M U L T I VA R I A T E tion of the F-ratio in which variability due to

S TAT I S T I C S regression (i.e., to the relationship between birth

Multivariate analyses are invariably done by weight and maternal age) is contrasted with resid-

computer because computations are complex. To ual variability. The value of F (15.252) with 1 and

illustrate computer analyses for two multivariate 28 df is signiﬁcant at the .001 level, consistent with

techniques, we return to the example described in information from Figure 20-6.

Chapters 19 and 20, involving the implementation The actual regression equation is presented in

of a special prenatal intervention for low-income the next panel (Coefﬁcients). If we wanted to pre-

young women. Data for these examples were pre- dict new values of birth weight based on maternal

sented in Table 19-8. age at birth, the equation would be:

Birth weight (3.119 age) 48.040

Example of Multiple Regression

That is, predicted birth weight values would be

In Chapter 20, we tested the hypothesis that older equal to the regression coefﬁcient (b 3.119)

mothers in the sample had infants with higher birth times values of maternal age (X), plus the value of

weights than younger mothers, using Pearson’s r. the intercept constant (a 48.040).

The calculated value of r was .594, which was In regression, a statistical test can be per-

highly signiﬁcant, thereby supporting the research formed for each regression coefﬁcient by dividing

hypothesis. b by the standard error for b; this is shown here

Suppose now that we want to test whether we

can signiﬁcantly improve our ability to predict

infant birth weight by adding two predictor vari- *The adjusted R2 of .330 is the R2 after it has been adjusted to

reﬂect more closely the goodness of ﬁt of the regression model

ables in a multiple regression analysis: whether the in the population, through a formula that involves sample size

mother smoked while pregnant, and number of and number of independent variables.

Regression

Variables Entered/Removeda

Variables Variables

Model Entered Removed Method

1 Stepwise

(Criteria:

Probability

-of-F-to-en

Mother’s ter <=

age .050,

Probability

-of-F-to-re

move >=

.100)

Model Summary

Adjusted R Std. Error of

Model R R Square Square the Estimate

ANOVAb

Sum of

Model Squares df Mean Square F Sig.

Residual 2253.017 28 80.465

Total 3480.300 29

b. Dependent Variable: Infant birth weight in ounces

Coefﬁcientsa

Standardi

zed

Unstandardized Coefﬁcients

Coefﬁcients

Mother’s age 3.119 .799 .594 3.905 .001

Excluded Variablesb

Colinearif

Partial y Statistics

Model Beta In t Sig. Correlation Tolerance

Smoking status .072a .446 .659 .086 .910

b. Dependent Variable: infact birth weight in ounces

FIGURE 21.8 Statistical Package for the Social Sciences (SPSS) printout for a stepwise multiple regression.

539

540 ■ PART 5 The Analysis of Research Data

under the columns t and Sig. The value of t for the the next panel (Tests of between-subjects

age coefﬁcient is 3.905, which is signiﬁcant at the effects), we see that the F-value for the covariate

.001 level. AGE is 24.358, significant at the .000 level (i.e.,

The ﬁnal panel (Excluded Variables) shows the beyond the .001 level). After controlling for

two predictors not yet in the equation, number of AGE, the F-value for the independent variable

prior pregnancies and smoking status. The ﬁgure GROUP is 8.719, which is significant at the .006

shows that the t values associated with the regres- level. In other words, once AGE is controlled, the

sion coefﬁcients for the two predictors are both research hypothesis about experimental versus

nonsigniﬁcant (p .915 and .659, respectively), control differences is supported rather than

once variation due to maternal age is taken into rejected. The total amount of variability

account. Because of this fact, the stepwise regres- explained with the two variables (1777.228),

sion ends at this point, without adding new predic- when contrasted with residual variability

tors. Given the signiﬁcance criterion of .05 (shown (1703.072), is also significant (F 14.088,

in the top panel of “probability of F to p .000). The multiple R2 for predicting birth

enter .050”), neither of the two predictor vari- weight, based on both AGE and GROUP, is

ables would have signiﬁcantly added to the predic- .511—substantially more than the R2 between

tion of birth weight, over and above what was maternal age and birth weight alone (.352).

already achieved with maternal age. The next panel shows the overall mean

It is, of course, possible to force the two (104.70) for the sample; the standard error (1.45);

additional predictors into the regression equation and the 95% conﬁdence interval for estimating the

by doing a simultaneous regression. When this is population mean (CI 101.725 to 107.675). The

done (not shown in the table), the value of R bottom panel shows the group means after they are

increases from .594 to .598, a negligible and non- adjusted for maternal age (i.e., the results of an

significant increase. Thus, we conclude that we MCA). The original, unadjusted means for the

cannot improve our ability to predict infant birth experimental and control groups were 107.53 and

weight with the two additional predictors 101.87, respectively (Figure 20-5). After adjusting

available. for maternal age, however, the experimental mean

is 109.08, and the control mean is 100.32, a much

more sizable difference.

Example of Analysis of Covariance

In Chapter 20, we tested the hypothesis that infants

in the experimental group have higher birth

GUIDE TO MULTIVARIATE

weights than infants in the control group, using a t-

STATISTICAL TESTS

test. The computer calculated t to be 1.44, which

was nonsignificant with 28 df. The research With the widespread availability of computers and

hypothesis was therefore rejected. user-friendly programs for performing statistical

Through ANCOVA, we can test the same analysis, multivariate statistics have become

hypothesis controlling for maternal age, which, increasingly common in nursing studies. As with

as we have just seen, is significantly correlated bivariate statistical tests, the selection of a multi-

with birth weight. Figure 21-9 presents the print- variate procedure depends on several factors,

out for ANCOVA for this analysis, with birth including the nature of the research question and

weight as the dependent variable, maternal age the measurement level of the variables. Table 21-8

(AGE) as the covariate, and GROUP (experimen- summarizes some of the major features of several

tal versus control) as the independent variable. of the multivariate statistics discussed in this chap-

The first panel shows that the group variable ter and thus serves as an aid to the selection of an

involves 15 experimentals and 15 controls. In appropriate procedure.

CHAPTER 21 Analyzing Quantitative Data: Multivariate Statistics ■ 541

Between-Subjects Factors

Value Label N

Treatment 1 Experimenta 15

group l

2 Control 15

Dependent Variable: Infant birth weight in ounces

Source of Squares df Mean Square F Sig.

Intercept 572.734 1 572.734 9.080 .006

AGE 1536.395 1 1536.395 24.358 .000

GROUP 549.945 1 549.945 8.719 .006

Error 1703.072 27 63.077

Total 332343.000 30

Corrected Total 3480.300 29

1. Grand Mean

Dependent Variable: Infant birth weight in ounces

95% Conﬁdence Interval

Mean Std. Error Lower Bound Upper Bound

104.700a 1.450 101.725 107.675

2. Treatment group

Dependent Variable: Infant birth weight in ounces

95% Conﬁdence Interval

Treatment group Mean Std. Error Lower Bound Upper Bound

Experimental 109.080a 2.074 104.824 113.337

Control 100.320a 2.074 96.063 104.576

FIGURE 21.9 Printout for ANCOVA analysis.

TABLE 21.8 Guide to Selected Multivariate Analyses

OF VARIABLES* VARIABLES†

correlation between 2 IVs and 1

DV; to predict a DV

from 2 IVs

(ANCOVA) between the means of 2

groups, while controlling

for 1 covariate

of variance between the means of

(MANOVA) 2groups for 2 DVs

simultaneously

of covariance between the means of 2

(MANCOVA) groups for 2 DVs simul-

taneously, while controlling

for 1 covariate

between 2 IVs and 1

DV; to predict group

membership; to classify

cases into groups

between 2 sets of variables

sionality and structure of

a set of variables

between 2 IVs and 1

DV; to predict the

probability of on event;

to estimate relative risk

†Measurement levels: N, nominal; I, interval; R, ratio.

CHAPTER 21 Analyzing Quantitative Data: Multivariate Statistics ■ 543

trolling a number of background characteristics strength-

Johnson and her colleagues (1999) conducted an evalua- ened the internal validity of the study.

tion of a nurse-delivered smoking cessation intervention

for hospitalized cardiac disease patients and used many

multivariate procedures. They used a nonequivalent con- S U M M A RY P O I N T S

trol group before—after design, using smokers admitted

• Multivariate statistical procedures are

to two inpatient cardiac units in a large Canadian hospi-

increasingly being used in nursing research to

tal. Patients in one unit received the intervention, which

involved two in-hospital contacts followed by 3 months untangle complex relationships among three or

of telephone support. Patients in the other unit received more variables.

no special intervention. • Simple linear regression makes predictions

The main outcome measures in the study were (1) about the values of one variable based on values

self-reported smoking status 6 months after enroll- of a second variable. Multiple regression is a

ment, and (2) self-reported smoking cessation self-effi- method of predicting a continuous dependent

cacy, measured using the Smoking Abstinence Self- variable on the basis of two or more indepen-

Efficacy Scale. When factor analyzed, this 20-item dent variables.

Likert scale revealed three stable factors, which were • The multiple correlation coefﬁcient (R) can be

used to form three subscales: a positive/social sub-

squared (R2) to estimate the proportion of vari-

scale, a negative/affective subscale, and a habit/addic-

tive subscale.

ability in the dependent variable accounted for

Using bivariate tests (t-tests and chi-squared tests), by the independent variables.

the researchers compared baseline characteristics of the • Simultaneous multiple regression enters all

experimental and comparison group in terms of age, gen- predictor variables into the regression equation at

der, marital status, education, and other demographic the same time. Hierarchical multiple regres-

characteristics. The two groups were found to be compa- sion enters predictors into the equation in a series

rable in most respects, but the experimental group was of steps controlled by researchers. Stepwise

signiﬁcantly more afﬂuent and signiﬁcantly less likely to multiple regression enters predictors in steps

have been admitted for cardiac surgery. The groups had using a statistical criterion for order of entry.

similar smoking histories. • Analysis of covariance (ANCOVA) extends

Six months after enrollment in the study, 30.8% of

ANOVA by removing the effect of extrane-

control subjects and 46.0% of experimental subjects

reported they were not smoking. A chi-squared test indi-

ous variables (covariates) before testing

cated that, although sizable, this difference was not statis- whether mean group differences are statistically

tically signiﬁcant. The researchers then conducted a logis- signiﬁcant.

tic regression to predict smoking abstinence based on • Multiple classiﬁcation analysis (MCA) is a

group assignment, while controlling for important covari- supplement to ANCOVA that yields information

ates (e.g., income, surgical status, prior smoking history). about the adjusted means of groups on a

They found that those in the control group were about dependent variable, after removing the effect of

three times as likely as those in the experimental group to covariates.

resume smoking, net of background factors (OR 3.18), • Factor analysis is used to reduce a large set of

which was signiﬁcant at the .05 level. variables into a smaller set of underlying dimen-

Multiple regression was used to predict scores on

sions, called factors. Mathematically, each fac-

the three subscales of the self-efﬁcacy scale. These

analyses revealed that, with background variables con-

tor is a linear combination of variables in a data

trolled, being in the treatment group was a signiﬁcant matrix.

predictor of scores on the positive/social and habit/addic- • The ﬁrst phase of factor analysis (factor extrac-

tive subscales. The total amount of variance accounted tion) identiﬁes clusters of variables with a high

for in these two regression analyses was 33% and 39%, degree of communality and condenses a larger

respectively. Although the researchers used a quasi- set of variables into a smaller number of factors.

544 ■ PART 5 The Analysis of Research Data

• The second phase of factor analysis involves • LISREL proceeds in two sequential phases: the

factor rotation, which enhances the inter- measurement model phase and the structural

pretability of the factors by aligning variables equations modeling phase. The measurement

more distinctly with a particular factor. model stipulates the hypothesized relationship

• Factor loadings shown in a rotated factor between latent variables and manifest variables;

matrix can be examined to identify and name if the measurement model ﬁts the data, the

the underlying dimensionality of the original set hypothesized causal model is then tested.

of variables and to compute factor scores. • Most multivariate procedures described in the

• Discriminant analysis is used to make predic- chapter are based on least-squares estimation,

tions about dependent variables that are cate- but LISREL is based on a different estimation

gorical (i.e., predictions about membership in procedure known as maximum likelihood;

groups) on the basis of two or more predictor maximum likelihood estimates the parameters

variables. most likely to have generated observed data.

• Canonical correlation analyzes the relation- • Logistic regression, another multivariate proce-

ship between two or more independent and two dure based on maximum likelihood estimation,

or more dependent variables; it yields a canoni- is used to predict categorical dependent vari-

cal correlation coefﬁcient, RC. ables. A useful feature of this technique is that it

• Multivariate analysis of variance (MANOVA) yields an odds ratio that is an index of relative

is the extension of ANOVA to situations in risk, that is, the risk of an outcome occurring

which there is more than one dependent variable. given one condition, versus the risk of it occur-

• Causal modeling involves the development and ring given a different condition.

testing of a hypothesized causal explanation of • Survival analysis and other related event histo-

a phenomenon. ry methods are used when the dependent vari-

• Path analysis, a regression-based method for able of interest is a time interval (e.g., time from

testing causal models, involves the preparation onset of a disease to death).

of a path diagram that stipulates hypothesized

causal links among variables. Path analysis tests STUDY ACTIVITIES

recursive models, that is, ones in which causa-

Chapter 21 of the accompanying Study Guide for

tion is presumed to be unidirectional.

Nursing Research: Principles and Methods, 7th

• Path analysis applies ordinary least-squares

ed., offers various exercises and study suggestions

regression to a series of structural equations,

for reinforcing the concepts presented in this chap-

resulting in a series of path coefﬁcients. The

ter. In addition, the following study questions can

path coefﬁcients represent weights associated

be addressed:

with a causal path, in standard deviation units.

• Linear structural relations analysis (LIS- 1. Refer to Figure 21-1. What would the value of

REL), another approach to causal modeling, Y be for the following X values: 8, 1, 3, 6?

does not have as many assumptions and restric- 2. A researcher has examined the relationship

tions as path analysis. LISREL can accommo- between preventive health care attitudes on the

date measurement errors, nonrecursive models one hand and the person’s educational level, age,

that allow for reciprocal causal paths, and corre- and gender on the other. The multiple correla-

lated errors. tion coefﬁcient is .62. Explain the meaning of

• LISREL can analyze causal models involving this statistic. How much variation in attitudinal

latent variables, which are not directly scores is explained by the three predictors? How

measured but which are captured by two or more much is unexplained? What other variables

manifest variables (i.e., measured variables). might improve the power of the prediction?

CHAPTER 21 Analyzing Quantitative Data: Multivariate Statistics ■ 545

3. Using power analysis, determine the sample Hair, J. F., Anderson, R. E., Tatham, R. L., & Black, W.

size needed to achieve power .80 for (1998). Multivariate data analysis (5th ed.). Upper

.05, when: Saddle River, NJ: Prentice-Hall.

a. estimated R2 .15, k 5 Harrell, F. E. (2001). Regression modeling strategies:

b. estimated R2 .33, k 10 With applications to linear models, logistic regression,

and survival analysis. New York: Springer-Verlag.

4. Which multivariate statistical procedures

Holt, F., Merwin, E., & Stern S. (1996). Alternative sta-

would you recommend using in the following

tistical methods to use with survival data. Nursing

situations: Research, 45, 345–349.

a. A researcher wants to test the effectiveness Hougaard, P. (2000). Analysis of multivariate survival

of a nursing intervention for reducing stress data. New York: Springer-Verlag.

levels among surgical patients, using an Jaccard, J., & Becker, M. A. (2001). Statistics for the

experimental group of patients from one behavioral sciences (4th ed.). Belmont, CA:

hospital and a control group from another Wadsworth.

hospital. Jöreskog, K. G., & Sörbom, D. (1997). LISREL 8: User’s

b. A researcher wants to predict which college reference guide. Mooresville, IN: Scientiﬁc Software

students are at risk of a sexually transmitted International.

disease using background information such Kelloway, E. K. (1998). Using LISREL for structural

as gender, academic grades, socioeconomic equation modeling: A researcher’s guide. Thousand

status, religion, and attitudes toward sexual Oaks, CA: Sage.

activity. Kim, J., Kaye, J., & Wright L. K. (2001). Moderating

c. A researcher wants to test the effects of and mediating effects in causal models. Issues in

Mental Health Nursing, 22, 63–75.

three different diets on blood sugar levels

Knapp, T. R. (1994). Regression analysis: What to

and blood pH.

report. Nursing Research, 43, 187–189.

Knapp, T. R., & Campbell-Heider, N. (1989). Numbers

SUGGESTED READINGS of observations and variables in multivariate analy-

ses. Western Journal of Nursing Research, 11,

Methodologic References 634–641.

Aaronson, L. S. (1989). A cautionary note on the use of Lee, E. T. (1992). Statistical methods for survival data

stepwise regression. Nursing Research, 38, 309–311. analysis (2nd ed.). New York: John Wiley & Sons.

Aaronson, L. S., Frey, M., & Boyd, C. J. (1988). Mason-Hawkes, J. & Holm, K. (1989). Causal modeling:

Structural equation models and nursing research: Part A comparison of path analysis and LISREL. Nursing

II. Nursing Research, 37, 315–318. Research, 38, 312–314.

Blalock, H. M., Jr. (1972). Causal inferences in nonex- Musil, C. M., Jones, S. L., & Warner, C. D. (1998).

perimental research. New York: W. W. Norton. Structural equation modeling and its relationship to

Bollen, K., & Bollen, W. (1989). Structural equations multiple regression and factor analysis. Research in

with latent variables. New York: John Wiley. Nursing & Health, 21, 271–281.

Boyd, C. J., Frey, M. A., & Aaronson, L. S. (1988). Pedhazur, E. J. (1997). Multiple regression in behavioral

Structural equation models and nursing research: Part research (3rd ed.). New York: Harcourt College

I. Nursing Research, 37, 249–253. Publishers.

Cohen, J., & Cohen, P. (1984). Applied multiple regres- Polit, D. F. (1996). Data analysis and statistics for nurs-

sion: Correlation analysis for behavioral sciences ing research. Stamford, CT: Appleton & Lange.

(2nd ed.). Hillsdale, NJ: Lawrence Erlbaum. Tabachnick, B. G., & Fidell, L. S. (2000). Using multi-

Goodwin, L. D. (1984). Increasing efficiency and variate statistics (4th ed.). New York: HarperCollins

precision of data analysis: Multivariate vs. univariate College Publishers.

statistical techniques. Nursing Research, 33, Weisberg, S. (1985). Applied linear regression (2nd ed.).

247–249. New York: John Wiley & Sons.

546 ■ PART 5 The Analysis of Research Data

Welkowitz, J., Ewen, R. B., & Cohen, J. (2000). Johnson, J. L., Budz, B., Mackay, M., & Miller, C.

Introductory statistics for the behavioral sciences (1999). Evaluation of a nurse-delivered smoking ces-

(5th ed.). New York: Harcourt College Publishers. sation intervention for hospitalized patients with car-

Wu, Y. B., & Slakter, M. J. (1989). Analysis of covariance diac disease. Heart & Lung, 28, 55–64.

in nursing research. Nursing Research, 38, 306–308. Lucas, J. A., Orshan, S. A., & Cook, F. (2000).

Determinants of health-promoting behavior among

women ages 65 and above living in the community.

Studies Cited in Chapter 21

Scholarly Inquiry for Nursing Practice, 14, 77–100.

Aminzadeh, F., & Edwards, N. (2000). Factors associated Mahon, N. E., & Yarcheski, A. (2001). Outcomes of

with cane use among community dwelling older depression in early adolescents. Western Journal of

adults. Public Health Nursing, 17, 474–483. Nursing Research, 23, 360–375.

Badger, T. A., & Collins-Joyce, P. (2000). Depression, McCarter-Spaulding, D. E. & Kearney, M. H. (2001).

psychosocial resources, and functional ability in Parenting self-efﬁcacy and perception of insufﬁcient

older adults. Clinical Nursing Research, 9, 238–255. breast milk. Journal of Obstetric, Gynecologic, and

Badr (Zahr), L. K. (2001). Quantitative and qualitative Neonatal Nursing, 30, 515–522.

predictors of development for low birth-weight Misener, T. R., & Cox, D. L. (2001). Development of the

infants of Latino background. Applied Nursing Misener Nurse Practitioner Job Satisfaction Scale.

Research, 14, 125–135. Journal of Nursing Measurement, 9, 91–108.

Berger, A. M., & Walker, S. N. (2001). An explanatory Nyamathi, A., Leake, B., Longshore, D., & Gelberg, L.

model of fatigue in women receiving adjuvant breast (2001). Reliability of homeless women’s reports:

cancer chemotherapy. Nursing Research, 50, 42–52. Concordance between hair assay and self report of

Cowan, M. J., Pike, K., & Budzynski, H. K. (2001). cocaine use. Nursing Research, 50, 165–171

Psychosocial nursing therapy following sudden car- Tarkka, M. T., Paunonen, M., & Laippala, P. (2000).

diac arrest: Impact on two-year survival. Nursing First-time mothers and child care when the child is 8

Research, 50, 68–76. months old. Journal of Advanced Nursing, 31, 20–26.

Fuller, B. F. (2000). Fluctuations in established infant pain

behaviors. Clinical Nursing Research, 9, 298–316.

- Numerical Method for engineers-chapter 19Загружено:Mrbudakbaek
- 확률통계답안지Загружено:likemun
- INTEGRATION OF IMPORTANCEPERFORMANCE ANALYSIS AND FUZZY DEMATELЗагружено:Anonymous Gl4IRRjzN
- Unit 5 ( CORRELATION AND REGRESSION )Загружено:Zara Nabilah
- ResearchЗагружено:ijayathunga
- Credit Expansion and Financial Instability: Evidence from Stock PricesЗагружено:mirando93
- Evaluating the Uncertainty of Polynomial Regression Models Using ExcelЗагружено:Satria Wijaya
- Writing an Honours Thesis - Joe Wolfe - UNSWЗагружено:rizkibizniz
- Analytical Delay Models for Signalized Intersections AkgungorЗагружено:Eka Swastika AlwiPutri
- summary of statsЗагружено:Nabhonil Basu Choudhury
- Do Clicks Count to Increase Student Achievement? Learner Analytics on a Large-Enrollment Hybrid Course - Sponsored furniture provided by Steelcase, Gold Partner (176641159)Загружено:EDUCAUSE
- ECO 231 Empirical Project Mets AttendanceЗагружено:mulcahk1
- 30 Correlation Main DRAFTЗагружено:Parthiban Rajendran
- Assignment 2Загружено:marc
- Chapter_01_Regression_5E(1).pdfЗагружено:khoold93
- Stats Ch 6 Final ReviewЗагружено:Ikequan Scott
- Analysis of Different Urban Forms in IstanbulЗагружено:Ismail Ibrahim Hajani
- SPSS_DescriptiveInferentialЗагружено:lekha1997
- 2. FORCASTINGЗагружено:Ucca Amanda
- 718849Загружено:carol
- LinearRegression-LectureNotesPublic.pdfЗагружено:Edmond Z
- Eva as Statistical ModelsЗагружено:leviandmeri
- L01Загружено:Kiju Jung
- jurnal Gerontik.pdfЗагружено:Andika Fahrurozi
- Data AnalysisЗагружено:Daniela Abisambra
- Jurnal SimЗагружено:ayyu purnama
- 2647-10536-1-PBЗагружено:Ayunda vidia
- Design of Experimentation for Cupola FurnaceЗагружено:International Journal of Innovations in Engineering and Science
- Farias Et Al 2018Загружено:JPDGL
- Inactivación Térmica de Listeria Monocytogenes en Productos Alimentarios LíquidosЗагружено:Marcela Martinez

- DILGU 1Загружено:mahadabrata21
- OUTPUT SPSS WORD.docxЗагружено:mahadabrata21
- 71-80_Manuskrip_Vevi_SЗагружено:Meita Isma
- 626-685-4-PBЗагружено:mahadabrata21
- Improving the Patient ExperienceЗагружено:mahadabrata21
- 605379Загружено:mahadabrata21
- FORMAT CVЗагружено:mahadabrata21
- EMI2014-215329Загружено:mahadabrata21
- 193Загружено:mahadabrata21
- Patient Safety-Modifying Processes to Eliminate Medical ErrorsЗагружено:mahadabrata21
- Hains 2010Загружено:mahadabrata21
- 4 Safe TransportЗагружено:Iqbal Tawakkal
- 2009 Jul TeamworkЗагружено:mahadabrata21
- 10.1.1.695.3351Загружено:mahadabrata21
- 2180-8305-1-PBЗагружено:Anonymous nEQNlgbYQC
- 44 WagnerC Quality 2006Загружено:mahadabrata21
- 28 4 FergusonЗагружено:mahadabrata21
- BUK RENI 3Загружено:mahadabrata21
- Code of Ethics en Vers 2014Загружено:Albert Einstein
- vol23.4-6Загружено:mahadabrata21
- Pub 1032621Загружено:mahadabrata21
- 0bdbd2eae62ce42a6323fa6cce65f0f71785Загружено:mahadabrata21
- 0903_03Загружено:mahadabrata21
- OJPM20120200006_73804781Загружено:mahadabrata21
- Vol7Issue1–Nursing Management in Saudi Arabia- Evaluating Strategies for Succession Planning 0d31Загружено:mahadabrata21
- Nursing_staff_teamwork_and_job_satisfaction.pdfЗагружено:Paryanto Hippii Solo

- Confidence Intervals for Intraclass CorrelationЗагружено:scjofyWFawlroa2r06YFVabfbaj
- Customer loyaltyЗагружено:debian_alex
- Seminar 5 Mini CaseЗагружено:bbaker2014
- BI Survey.pdfЗагружено:Tuti Aryanti Hastomo
- Giovannucci 2017 NIPSЗагружено:Andrea Giovannucci
- Statistic PROJECTЗагружено:Javed Bumlajiwala
- Navidi Ch07 4e Linear RegressionЗагружено:Amin Zaquan
- Sensation and Type of MusicЗагружено:gadir
- A guide to using EviewsЗагружено:Vishyata
- correlationЗагружено:Sanjid Elahi
- Stats Summary Sheet.pdfЗагружено:MbusoThabethe
- 2009 Corneal Refractive Power Estimation and Intraocular Lens Calculation After Hyperopic LASIKЗагружено:Hanna_R
- thesis pcЗагружено:divlion098117
- Syllabus B ComЗагружено:Prajapat Chetankumar
- Determination of Carcass and Body Fat Compositions of Grazing Crossbred Bulls UsingЗагружено:Olga Lucia Duque
- Effects influencing the compressive strength of a solid, fired clay brickЗагружено:kalmaster1
- Determination of the Optimal Dosage of Aluminum Sulfate in the CoagulationЗагружено:muaz
- The Use of StatisticsЗагружено:angelialoh0929
- vol2no2Загружено:Giorgos Staf
- 17_RubashviliЗагружено:DIta
- 69-81Загружено:Aman Khokhar
- Technetium HMPAO SPECT Study in Dementia with Lewy Bodies, Alzheimer's Disease and Idiopathic Parkinson's DiseaseЗагружено:Mihaela Toader
- Al2000Загружено:Asdrúbal Arteaga
- MA7155-Applied Probability and Statistics Question BankЗагружено:selvakrishnan_s
- c3Загружено:crwithanage
- mixedЗагружено:mphil.ramesh
- Statistics Primer.pdfЗагружено:sharmaegayangos
- PAST - Multivariate StatisticsЗагружено:Zahid_Sharif_8414
- Article 3.pdfЗагружено:JohnnyNewhouse
- Chap 012Загружено:BG Monty 1