Академический Документы
Профессиональный Документы
Культура Документы
Methods http://orm.sagepub.com/
A Critical Examination of Common Beliefs About Partial Least Squares Path Modeling
Mikko Rnkk and Joerg Evermann
Organizational Research Methods 2013 16: 425 originally published online 7 March 2013
DOI: 10.1177/1094428112474693
Published by:
http://www.sagepublications.com
On behalf of:
Additional services and information for Organizational Research Methods can be found at:
Subscriptions: http://orm.sagepub.com/subscriptions
Reprints: http://www.sagepub.com/journalsReprints.nav
Permissions: http://www.sagepub.com/journalsPermissions.nav
What is This?
Downloaded from orm.sagepub.com at NATIONAL CHUNG HSING UNIV on April 12, 2014
Article
Organizational Research Methods
16(3) 425-448
The Author(s) 2013
A Critical Examination of Reprints and permission:
sagepub.com/journalsPermissions.nav
Common Beliefs About DOI: 10.1177/1094428112474693
orm.sagepub.com
Partial Least Squares Path
Modeling
Abstract
Partial least squares path modeling (PLS) was developed in the 1960s and 1970s as a method for
predictive modeling. In the succeeding years, applied disciplines, including organizational and man-
agement research, have developed beliefs about the capabilities of PLS and its suitability for different
applications. On close examination, some of these beliefs prove to be unfounded and to bear little
correspondence to the actual capabilities of PLS. In this article, we critically examine several of these
commonly held beliefs. We describe their origins, and, using simple examples, we demonstrate that
many of these beliefs are not true. We conclude that the method is widely misunderstood, and our
results cast strong doubts on its effectiveness for building and testing theory in organizational research.
Keywords
partial least squares, statistical and methodological myths and urban legends, structural equation
modeling
Partial least squares path modeling (PLS) was developed in the 1960s and 1970s by Herman Wold
(cf. Joreskog & Wold, 1982) as an alternative to LISREL. However, Dijkstra (1983) soon proved
that if PLS is used as an estimator for structural equation models (SEMs), the parameter estimates
are both inconsistent and biased, leading to abandonment of further development. Nevertheless, PLS
experienced a renaissance in more applied disciplines with proponents such as Fornell and Book-
stein (1982) and later Chin (1998) and Hulland (1999), and recent publications in management
journals indicate that its use is increasing (Antonakis, Bendahan, Jacquart, & Lalive, 2010; Echam-
badi, Campbell, & Agarwal, 2006; Gruber, Heinemann, Brettel, & Hungeling, 2010; Hair, Sarstedt,
Pieper, & Ringle, 2012; Peng & Lai, 2012; Reinartz, Haenlein, & Henseler, 2009; Sosik, Kahai, &
1
Aalto University, Aalto, Finland
2
Memorial University of Newfoundland, St. Johns, Canada
Corresponding Author:
Mikko Ronkko, Aalto University School of Science, PO Box 15500, FI-00076 Aalto, Finland.
Email: mikko.ronkko@aalto.fi
Downloaded from orm.sagepub.com at NATIONAL CHUNG HSING UNIV on April 12, 2014
426 Organizational Research Methods 16(3)
Piovoso, 2009). Following Atinc, Simmering, and Kroll (2011) we reviewed four leading manage-
ment journals, Academy of Management Journal, Journal of Applied Psychology, Journal of
Management, and Strategic Management Journal, and found 27 studies that used PLS, which are
listed in Table 1; a third of these studies were published in the past 5 years, supporting the argument
that the use of PLS is becoming more common.
In contrast to its popularity in management research, PLS has been largely ignored in research
methods journals. For example, there are no articles in Organizational Research Methods addressing
the PLS method. In our review of other top research methods journals, we found one article about
PLS (Henseler & Chin, 2010) published in Structural Equation Modeling and one article (Dijkstra,
1983) in Journal of Econometrics. No articles about PLS were found in Psychological Methods,
Psychological Bulletin, or Econometrica.
The absence of articles on PLS in the research methods literature has led researchers in disci-
plines such as strategic management (Hulland, 1999), operations management (Peng & Lai,
2012), marketing (Hair, Ringle, & Sarstedt, 2011), group research (Sosik et al., 2009), and informa-
tion systems (Gefen, Rigdon, & Straub, 2011) to develop their own guidelines on how to perform
and evaluate PLS-based studies. We argue that most of these articles present an overly positive pic-
ture of the method, with some aggressively promoting the method as a silver bullet (Hair et al.,
2011), a success story (Vinzi, Chin, Henseler, & Wang, 2010), or as a method with genuine
advantages (Henseler & Sarstedt, 2013). However, many of these articles are not based on statis-
tical theory or simulation studies but are based on beliefs about the method that earlier, similar arti-
cles have presented, leading to the perpetuation of commonly held beliefs that have not been
demonstrated, that is, methodological myths and urban legends (Vandenberg, 2006).
This article addresses some of these beliefs and shows that they are incorrect or correct only with
strong qualifications. Given the growing popularity of PLS-based studies in management research,
this is a timely and important topic.
At this level of abstraction, PLS estimation is identical to estimation with OLS regression on
summed scales or factor scores. PLS differs from these methods in that the indicator weights o are
based on the estimated model and the sample data. The iterative weighting process begins by
approximating the latent variables as unweighted sums and then iteratively adjusting the indicator
weights and approximate latent variable scores Y~ in two steps, called inner and outer estimation,
until the weights and scores converge (Table 2).
Downloaded from orm.sagepub.com at NATIONAL CHUNG HSING UNIV on April 12, 2014
Table 1. Use of PLS in Selected Management Journals Ordered by Year (Academy of Management Journal, Journal of Applied Psychology, Journal of Management, Strategic Man-
agement Journal).
How PLS Was Used Relationship to SEM Capabilities of PLS
Summary of Description Null Hypothesis Simultaneous PLS Is Latent Distribution Models or Reduces Appropriate for
of the Method and Why Validation of Significance Eqs./Path Variable Free/Nonnormal Measurement Works Well With Early/Exploratory
Article It Was Used Measurement Testing PLS Is SEM Analysis Analysis Data Error Small Samples Research
Downloaded from orm.sagepub.com at NATIONAL CHUNG HSING UNIV on April 12, 2014
generation method
that examines the
relationships in the
theoretical model
simultaneously. PLS
was used because it is
well suited for
assessing predictive
relationships.
(continued)
427
Table 1. (continued)
428
How PLS Was Used Relationship to SEM Capabilities of PLS
Summary of Description Null Hypothesis Simultaneous PLS Is Latent Distribution Models or Reduces Appropriate for
of the Method and Why Validation of Significance Eqs./Path Variable Free/Nonnormal Measurement Works Well With Early/Exploratory
Article It Was Used Measurement Testing PLS Is SEM Analysis Analysis Data Error Small Samples Research
Downloaded from orm.sagepub.com at NATIONAL CHUNG HSING UNIV on April 12, 2014
multivariate technique
that is ideal for testing
structural models
with latent variables.
Delios and Beamish, No description of the
1999 method given. PLS
was used because the
objective of the study
was prediction of the
dependent variable.
(continued)
Table 1. (continued)
Summary of Description Null Hypothesis Simultaneous PLS Is Latent Distribution Models or Reduces Appropriate for
of the Method and Why Validation of Significance Eqs./Path Variable Free/Nonnormal Measurement Works Well With Early/Exploratory
Article It Was Used Measurement Testing PLS Is SEM Analysis Analysis Data Error Small Samples Research
Downloaded from orm.sagepub.com at NATIONAL CHUNG HSING UNIV on April 12, 2014
Fladmoe-Lindquist, technique that can
2002 model both reflective
and formative
indicators and is well
suited to assessing
predictive relations.
Bass, Avolio, Jung, and PLS is a structural
Berson, 2003 equation modeling
technique that has less
stringent assumptions
about the data.
429
(continued)
430
Table 1. (continued)
Summary of Description Null Hypothesis Simultaneous PLS Is Latent Distribution Models or Reduces Appropriate for
of the Method and Why Validation of Significance Eqs./Path Variable Free/Nonnormal Measurement Works Well With Early/Exploratory
Article It Was Used Measurement Testing PLS Is SEM Analysis Analysis Data Error Small Samples Research
Downloaded from orm.sagepub.com at NATIONAL CHUNG HSING UNIV on April 12, 2014
exploratory nature of
the research.
Gruber et al., 2010 PLS is a variance-based
structural equation
modeling technique. It
was used for its ability
to accommodate for-
mative indicators.
(continued)
Table 1. (continued)
Summary of Description Null Hypothesis Simultaneous PLS Is Latent Distribution Models or Reduces Appropriate for
of the Method and Why Validation of Significance Eqs./Path Variable Free/Nonnormal Measurement Works Well With Early/Exploratory
Article It Was Used Measurement Testing PLS Is SEM Analysis Analysis Data Error Small Samples Research
Downloaded from orm.sagepub.com at NATIONAL CHUNG HSING UNIV on April 12, 2014
studies
Percentage of surveyed 93 93 78 30 41 63 15 37 33
studies
431
432 Organizational Research Methods 16(3)
Stage 1 Iterative estimation of weights and LV scores; starting at Step 4 with arbitrary weights, repeat Steps 1
to 4 until convergence is achieved
sign covYj ; Yi if Yj and Yi are adjacent
1. Inner weights (centroid weighting scheme)a vji
P 0 otherwise
2. Inside approximation Y~j vji Yi
i
3. Outer weights; estimate okj in Equation 4 (Mode A) or Equation 5 (Mode B) using ordinary least
squares regression P
4. Outside approximation Yej fj o ~ kj ykj
kj
Stage 2 Estimation of path and loading coefficients by ordinary least squares regression from Equation 1 (path
coefficients) and Equation 2 (loadings coefficients) where the latent variables Zj are replaced by their
estimates Yj
Stage 3 Estimation of the means of latent and manifest variables as weighted sums
Note: LV latent variable; PLS partial least squares.
a
Other weighting schemes are path and factor weighting. However, the centroid weighting scheme is most frequently used
and is the default method in much PLS software.
During the inner estimation step, new latent variable score approximations are calculated as
weighted sums of adjacent latent variable score approximations, that is, of latent variables related
to the focal variable by regression relationships. During the outer estimation step, new indicator
weights o~ are calculated in either of two ways. In Mode A estimation, the manifest variables y are
regressed on the approximations Y~:
The new indicator weights are then used to estimate new latent variable score approximations for the
following iteration of inner estimation. The basic PLS algorithm is shown in Table 2. The only
difference between PLS and OLS is the different method of indicator weighting.
Myth 1: PLS Has Advantages Over Traditional Methods Because It Is an SEM Estimator
Almost every article in Table 1 and all of the guidelines on PLS present it as an SEM method, with
some even emphasizing its differences from OLS on summed scales or factor scores (e.g., Gefen
et al., 2011). This characterization cannot be found in the original articles on PLS (e.g., Wold,
Downloaded from orm.sagepub.com at NATIONAL CHUNG HSING UNIV on April 12, 2014
Ronkko and Evermann 433
A B
a1
a 2
a
3
b 1
b2
b
3
a1 a2 a3 b1 b2 b3
a 1
a 2
a
3
b 1
b 2
b 3
1985b). Rather, it is attributable to a widely cited article by Fornell and Bookstein (1982). Never-
theless, even the original characterization as a latent variable modeling technique is misleading.
In contrast to claims by many studies in Table 1, PLS does not estimate path models with latent vari-
ables, but with composites, and instead of using path analysis with simultaneous equations, PLS uses
separate OLS regressions. Thus, it is conceptually closer to OLS regressions on summed scales or
factor scores than to covariance structure analysis.
Although PLS can technically be argued to be an SEM estimator, so can OLS regression with
summed scales or factor scores: Both fit the definition of the term estimator (Lehmann & Casella,
1998, p. 4) because they provide some estimates of model parameters. This, however, does not mean
that PLS or OLS are good estimators in the sense of being consistent and unbiased. An estimator is
consistent if its estimates converge to the population value as the sample size increases. It is
unbiased if the mean of repeated estimates using samples drawn from the same population
approaches the population value as the number of samples increases. PLS has been shown to have
neither of these properties. In fact, Wold (e.g., 1985b) is quite clear that PLS estimates are incon-
sistent, Dijkstra (1983) presented a proof of this, and more recent PLS literature (e.g., Chin,
1998) also readily acknowledges that PLS estimates are biased.
In addition to providing inconsistent and biased estimates, the lack of an overidentification test is
another disadvantage of PLS over SEM. Paths between variables in a simultaneous equations model
can be constrained to zero. Such constraints make the model overidentified and allow testing whether
the constrained model fits the data. Thus, this overidentification test can be used to rule out endogene-
ity (unmodeled dependencies) that would otherwise cause inconsistency of estimates (Antonakis et al.,
2010). Overidentification tests also allow the researcher to rule out alternative causes, which is a key
step in testing a model causally (Antonakis et al., 2010; Bollen, 1989, chap. 3). Because no overiden-
tification test is available for PLS, PLS cannot test a model causally, but is limited to estimating sta-
tistical associations.
Despite the evidence, many of the reviewed articles argue that PLS supports causal modeling
(Birkinshaw, Hood, & Jonsson, 1998; Cool, Dierickx, & Jemison, 1989; Delios & Beamish,
1999; Johansson & Yip, 1994; Shea & Howell, 2000), and some researchers have the misconception
that using PLS avoids inconsistent and biased parameter estimates. For example, Birkinshaw,
Morrison, and Hulland (1995) write that to avoid obtaining biased and inconsistent parameter esti-
mates for these equations, the [hypothesized model] must be analyzed using a multivariate estima-
tion technique such as two-stage least squares or PLS (p. 647). However, of these two estimators,
only two-stage least squares is consistent.
In summary, although the argument that PLS is an SEM estimator is technically true, it is as cor-
rect to state that OLS regression is an SEM estimator. The lack of unbiasedness and consistency
means that both methods will provide erroneous estimates. This, and the lack of an
Downloaded from orm.sagepub.com at NATIONAL CHUNG HSING UNIV on April 12, 2014
434 Organizational Research Methods 16(3)
overidentification test, means that any potential advantage that might be obtained by specifying a
research problem as an SEM model is lost (Antonakis et al., 2010; McDonald, 1996). The claim that
PLS provides advantages because it is an SEM method is a methodological myth, and the current
practice of labeling PLS as an SEM method, although correct in a strict technical sense, is very
misleading.
In the two-construct example, under the default PLS assumption of standardized latent variables and
indicators, the correlation between two indicators is the sum of the correlation caused by common
antecedents and error correlation. Using covariance algebra one can write the correlation between
indicators as follows:
rai bj pai blpbj Eai bj : Equation 7
Substituting Equation 7 into Equation 6 and rearranging the terms yields
X
3 X
3
ra1 B~ pa1 b obi pbi obi Ea1 bi : Equation 8
i1 i1
This equation shows that although the indicator weighting system takes the indicator reliability into
account, it is highly sensitive to the correlated errors ea1 bi in the second sum of Equation 8, contrary
to recent claims (Gefen et al., 2011). This is problematic because, as a result of sampling variation,
the errors are never exactly uncorrelated, even for correct models. When the latent variables are not
correlated (b 0), the effect is more pronounced because the first term in Equation 8 vanishes and
only the effect of correlated errors remains.
With PLS Mode B, the indicator weights are defined based on regressions of the latent variable
estimate on their indicators. Hence, they are also affected by the correlations between indicators of
the same construct. In this regression, a set of highly reliable indicators is highly collinear, resulting
in suppression effects and instability of the regression coefficients and the resulting indicator
weights.
Downloaded from orm.sagepub.com at NATIONAL CHUNG HSING UNIV on April 12, 2014
Ronkko and Evermann 435
30
PLS Mode A
PLS Mode B
Sum scales
25
25
20
20
Density
Density
15
15
10
10
5
5
0
0
Figure 2. Distribution of reliability for partial least squares (PLS) Mode A, PLS Mode B, and summed scales in
the two-construct model over 500 replications.
To illustrate these effects numerically for our two-construct example model, we simulated
data for two conditions, one with no effect between constructs (b 0) and one with a moderate
size effect (b 0:3). We set the factor loadings to 0.6, 0.7, and 0.8. Similar loadings have been
used in prior studies (Goodhue, Lewis, & Thompson, 2012; Reinartz et al., 2009). We used 100
observations as the sample size, which is fairly typical for a PLS study. We performed 500
replications for both experimental conditions. The models were estimated using the plspm
package for R (Sanchez & Trinchera, 2012). We generated the indicator data by first generating
a sample of construct true scores and then used these to generate the indicator data. This
enabled us to calculate the reliability of the construct estimates directly as a squared correlation
between the true scores and estimated latent variable values. Figure 2 compares the reliability
of PLS Mode A and PLS Mode B. We also included summed scales as it is the simplest way to
construct composite variables that any other method should exceed to be considered useful
(McDonald, 1996).
Figure 2 shows that for both PLS modes, summed scales provide better (more reliable) construct
scores, and PLS Mode B performs substantially worse than PLS Mode A. The reason for this lower
reliability can be seen in Table 3, which shows that, although on average more reliable indicators are
weighted more highly, the effect of random correlations causes large variance in the weights, so that
any individual replication is very unlikely to have weights even close to an optimal combination. In
addition, the collinearity and suppression effects for Mode B estimation are seen clearly in the higher
standard deviation of the weights. The lower means of Mode B weights are a result of occasional
negative weights caused by collinearity.
We now show how the weighting affects the regression estimates between the latent variables.
The standardized path estimate (b) ^ between composites A and B is equal to their correlation:
!
X3 X
3
^
b cor oa ai ; wb bj : Equation 9
i j
i1 j1
In PLS, the weights are chosen so that the construct estimates are standardized. Thus, the correlation
is equal to the covariance, and one can write the covariance of sums as
Downloaded from orm.sagepub.com at NATIONAL CHUNG HSING UNIV on April 12, 2014
436 Organizational Research Methods 16(3)
Table 3. Means (standard deviations) of Weight for Indicator a1 in the Example Over 500 Replications.
Weight of a1 (population loading 0.6, reliability 0.36) 0.27 (0.40) 0.09 (0.66) 0.35 (0.18) 0.20 (0.50)
Weight of a2 (population loading 0.7, reliability 0.49) 0.35 (0.36) 0.17 (0.70) 0.39 (0.17) 0.24 (0.53)
Weight of a3 (population loading 0.8, reliability 0.64) 0.36 (0.36) 0.20 (0.73) 0.45 (0.16) 0.45 (0.52)
X
3 X
3
^
b oai obj rai bj : Equation 10
i1 j1
Substituting Equation 7 into this and rearranging the terms yields the following:
3 X
X 3 3 X
X 3
^b
b oai obj pai pbj oai obj Eai bj : Equation 11
i1 j1 i1 j1
Equation 11 shows that not only the indicator weights but also the path coefficient estimates are
affected by the correlated errors. The bias caused by unmodeled correlations is well known
(Zimmerman & Williams, 1977), but the joint effect that causes this bias to be amplified by PLS
has not been documented in prior literature. This is illustrated in Figure 3, which shows that the para-
meter estimates obtained from PLS are strongly biased away from zero, whereas no such bias exists
when the model is estimated with regression on summed scales or SEM.3 This effect is especially
pronounced when b 0 where both Mode A and Mode B estimation produce a bimodal distribution
of estimates, symmetrical around 0. These artificially inflated parameter estimates can lead to arti-
ficial inflation of statistical significance and incorrect inference.
In contrast to the unsubstantiated claims that PLS reduces the effect of measurement error, we
have shown that the indicator weights are strongly affected by error correlations and even the small
chance correlations caused by sampling variation are sufficient to affect the weights, resulting in
lower reliability composites than even simple summed scales.
The options available for reducing the effect of measurement error with composite variables are
limited because any linear composite of indicators that contain error will also be contaminated with
error. Random error in the composites causes attenuation of bivariate correlations resulting in bias in
the regression estimates. Although it is possible to apply a correction and then use the disattenuated
correlations in regression analysis, ML estimation of SEM models has superseded this approach
(Cohen, Cohen, West, & Aiken, 2003, pp. 38-39, 473-474).
Downloaded from orm.sagepub.com at NATIONAL CHUNG HSING UNIV on April 12, 2014
Ronkko and Evermann 437
PLS Mode A
5
PLS Mode B
Sum scales
SEM
4
3
Density
3
Density
2
2
1
1
0
0
0.6 0.4 0.2 0.0 0.2 0.4 0.6 0.4 0.2 0.0 0.2 0.4 0.6 0.8
Parameter estimate Parameter estimate
Figure 3. Distribution of parameter estimates for partial least squares (PLS) Mode A, PLS Mode B, summed
scales, and structural equation modeling (SEM) in the two-construct model, 500 replications.
composite reliability (CR) metric and the average variance extracted (AVE) statistic, which are both
more or less directly based on factor loading estimates. In addition to these, a family of goodness-of-
fit (GoF) indices exists that are calculated on the basis of the endogenous variable R2 values and
indicator communalities (cf. Henseler & Sarstedt, 2013). There are also various standardized root
mean squares of different model residuals (SRMR) that Lohmoller (1989) proposed to be used for
model assessment, but we are not aware of any applications of these criteria.
Because the use of GoF indices was recently comprehensively and convincingly debunked by
Henseler and Sarstedt (2013) and because of the obscurity of the SRMR indices, we focus on the AVE
and CR statistics, which are commonly used in the studies that we reviewed. Aguirre-Urreta, Marakas,
and Ellis (2013) showed that the CR indices are severely biased estimate of reliability of the com-
posites. Moreover, the same issues apply also to the AVE statistic. First, their definitions do not
include information about the indicator weights (Fornell & Larcker, 1981, pp. 45-46) but assume
unweighted composites, which is intentionally violated in a PLS analysis. Second, they are based
on factor loadings, but PLS does not calculate factor loadings, but composite loadings (McDonald,
1996, p. 248). These are always higher than factor loadings as they also explain part of the error
variance, whereas a factor analysis explains only the common variance between the indicators.4
Consequently, the CR and AVE statistics are also overestimated.
To show the effect of model misspecifications on the AVE, CR, (relative) GoF, and (indicator)
SRMR statistics and to demonstrate that these heuristics are unable to reliably identify when the model
does not fit data, we calculated these model quality metrics for our two-construct model with 500 repli-
cations (Table 4). The estimated model is always our two-construct example, and we vary the popu-
lation model to make the estimated model misspecified (Misspecified A, B, C in Table 4). The last
column in Table 4 shows the percentage of models in each condition that a researcher would accept as
valid, based on commonly used cutoff values for the different heuristics.5 This percentage should be
high for the true model and low for the misspecified models. However, results show that this number is
high for the composite reliability heuristic across all conditions; researchers are likely to accept all mis-
specified models as valid. This number is low for the SRMR heuristic across all conditions; researchers
are likely to also reject the true model as invalid. The situation is even worse for the rGoF heuristic. By
this heuristic, researchers are likely to reject only the true model and are likely to accept only the
Downloaded from orm.sagepub.com at NATIONAL CHUNG HSING UNIV on April 12, 2014
438 Organizational Research Methods 16(3)
Table 4. Model Fit Indices for the Example Model Estimated With Data From Four Different Population
Models.
Composite reliability 0.875 0.940 0.960 98.0
A B
Average variance extracted 0.563 0.646 0.695 98.2
AVEhighest squared corr. 0.454 0.563 0.631 100
a1 a2 a3 b1 b2 b3
Relative goodness of fit 0.620 0.859 0.971 28.6
Std. root mean square residual 0.076 0.090 0.112 13.6
misspecified models. Finally, the two AVE-based heuristics appear to detect only the third misspeci-
fied model, making them unreliable indicators of misspecified models.
In contrast to these heuristics, the w2 test of model fit provides a statistically sound way of iden-
tifying misspecified models for medium and large sample sizes,6 and the field of psychometrics pro-
vides decades of guidance on how to validate measurement with factor analysis (cf. Nunnally, 1978).
If desired, factor loadings can be obtained with common factor analysis, and the reliability of a com-
posite with arbitrary weights can be estimated as described by Raykov (1997; see Aguirre-Urreta
et al., in press, for how to apply the procedure with PLS analysis). Because of these better alterna-
tives, the measurement model should never be evaluated based on the composite loadings produced
by PLS or any statistic derived from these. In summary, we conclude that the idea that PLS results
can be used to validate a measurement model is a myth.
Myth 4: PLS Can Be Used for Testing Null Hypotheses About Path Coefficients
Almost all of the articles listed in Table 1 use PLS for null hypothesis significance testing (NHST) of
path coefficients, a practice that can be traced back to Fornell and Bookstein (1982). Under NHST,
statistical inferences are made based on the p value, which is defined as the probability of obtaining
a value of a test statistic . . . as large as the one obtainedconditional on the null hypothesis being
true (Nickerson, 2000, p. 247). Thus, NHST relies on a known sampling distribution of the test
statistic when the null hypothesis of no effect holds.
Downloaded from orm.sagepub.com at NATIONAL CHUNG HSING UNIV on April 12, 2014
Ronkko and Evermann 439
5
5
4
4
3
3
Density
2
2
1
1
0
0
0.6 0.4 0.2 0.0 0.2 0.4 0.6 0.6 0.4 0.2 0.0 0.2 0.4 0.6
6
5
5
4
4
Density
3
2
2
1
1
0
0.4 0.2 0.0 0.2 0.4 0.6 0.8 0.4 0.2 0.0 0.2 0.4 0.6 0.8
Figure 4. Distribution of parameter estimates over 500 replications and distribution of 500 bootstrap esti-
mates for the first five replications for partial least squares (PLS) Mode A and PLS Mode B in the two-construct
model.
The current practice in PLS studies is to use bootstrapping to estimate the standard errors for the
parameter estimates, calculate the ratio of a parameter estimate to its standard error, and compare
this statistic to the t distribution to obtain the p value. The use of a t distribution assumes a normal
distribution of the underlying parameter estimates. However, as shown in Figure 3, the distribution is
not normal but of bimodal shape when the null hypothesis of no effect holds (b 0. The exact dis-
tribution of the path coefficients is unknown as it depends on the sampling distribution of the indi-
cator weights, which is unknown (Dijkstra, 1983). Thus, because NHST requires a test statistic with
a known sampling distribution, the PLS path estimates cannot be used in NHST.
The use of bootstrapped confidence intervals (Wood, 2005) is an alternative to NHST for statis-
tical inference. Even this approach can be problematic with PLS: Bootstrapping relies on the
assumption that the bootstrap estimates follow the same distribution as the original statistic, but this
is not always the case, leading to incorrect inference (cf. Bollen & Stine, 1992). Figure 4 shows the
Downloaded from orm.sagepub.com at NATIONAL CHUNG HSING UNIV on April 12, 2014
440 Organizational Research Methods 16(3)
parameter estimate distribution for 500 replications of the simulation estimates and, for the first 5 of
these replications, shows the bootstrap distribution of the parameters. It is clear that the bootstrapped
replications do not follow the original sampling distribution in our simple example model.7 Thus, we
conclude that bootstrapped confidence intervals of PLS estimates should not be used for making sta-
tistical inferences until further research is available to show under which conditions, if any, the boot-
strapped distribution follows the sampling distribution of the PLS parameter estimate.
Downloaded from orm.sagepub.com at NATIONAL CHUNG HSING UNIV on April 12, 2014
Sample size 100, beta = 0 Sample size 50, beta = 0 Sample size 25, beta = 0
PLS Mode A
PLS Mode B
2.5
Sum scales
3
1.5
SEM
2.0
Ronkko and Evermann
2
1.5
1.0
Density
1.0
1
0.5
0.5
0
0.0
0.0
0.6 0.4 0.2 0.0 0.2 0.4 0.6 0.6 0.4 0.2 0.0 0.2 0.4 0.6 0.6 0.4 0.2 0.0 0.2 0.4 0.6
Sample size 100, beta = 0.3 Sample size 50, beta = 0.3 Sample size 25, beta = 0.3
3.0
5
2.5
4
2.0
3
2
1.5
Density
2
1.0
Downloaded from orm.sagepub.com at NATIONAL CHUNG HSING UNIV on April 12, 2014
1
1
0.5
0
0.0
0.4 0.2 0.0 0.2 0.4 0.6 0.8 0.4 0.2 0.0 0.2 0.4 0.6 0.8 0.4 0.2 0.0 0.2 0.4 0.6 0.8
Parameter estimate Parameter estimate Parameter estimate
Figure 5. Distribution of parameter estimates for partial least squares (PLS) Mode A, PLS Mode B, summed scales, and structural equation modeling (SEM) with
different sample sizes in the two-construct model, 500 replications.
441
442 Organizational Research Methods 16(3)
the small-sample-size capabilities of PLS are a myth arising from ignoring the effect of sampling
error amplification and an inappropriate use of the t test for parameter significance.
The best remedy for small samples is to collect sufficient data to avoid the problem. SEM tech-
niques for small samples (e.g., Herzog & Boomsma, 2009) and for estimating sample size require-
ments (e.g., Lai & Kelley, 2011) are actively studied, but fundamental laws of probability limit what
can be accomplished. Because the sample size requirement is tied to the size of the model, one way
to reduce the required sample size is to reduce the number of indicators by parceling (Landis, Beal,
& Tesluk, 2000; Little, Cunningham, Shahar, & Widaman, 2002). Another potentially useful option
is to use the limited information 2SLS estimator instead of the full information ML.
Downloaded from orm.sagepub.com at NATIONAL CHUNG HSING UNIV on April 12, 2014
Ronkko and Evermann 443
analyses are actively developed (e.g., Asparouhov & Muthen, 2009). If, on the other hand, explora-
tory research refers to uncertainty about the model rather than the search for a model, we recommend
using the 2SLS estimator that is less sensitive to model misspecification than the ML estimator
(Bollen, Kirby, Curran, Paxton, & Chen, 2007).
Downloaded from orm.sagepub.com at NATIONAL CHUNG HSING UNIV on April 12, 2014
444 Organizational Research Methods 16(3)
regression with summed scales or factor scores. PLS may be useful for purely predictive analyses,
but we are not aware of any studies showing this to be the case either.
Funding
The author(s) received no financial support for the research, authorship, and/or publication of this article.
Notes
1. Lohmollers (1989) description of the PLS model, on which this section is based, uses the term residual (the
difference between observed and estimated value) instead of error (a theoretical source of random variation)
when discussing the population model. We have corrected the terminology in our description of the PLS
method.
2. There is a persisting and unsubstantiated belief originating from the article by Fornell and Bookstein (1982)
that Mode B would be appropriate for formative measurement. Because formative measurement is based on
assumptions that are very unlikely to ever hold (Edwards, 2011), we do not cover formative models in the
article, but use Mode B only to estimate reflective models as it was originally intended.
3. SEM models were estimated with Lavaan 0.4-14 (Rosseel, 2012).
4. For example, if the construct A in our example model has three indicators (a1, a2, and a3) each with a factor
P3
loading of 0.4, the covariance between one indicator and the composite is covaj ; ai 33
25. Because the
i1
3 P
P 3
99
variance of the composite is covai; aj 25, the correlation between an item and a composite is
i1 j1
p
covaj ;A
p 11
5 0.663, which overestimates the factor loading by 66%.
varA
5. Composite reliability > 0.7, average variance extracted > 0.5, AVEhighest squared correlation > 0 (Fornell
& Bookstein, 1982), relative goodness of fit > 0.9 (Gefen, Rigdon, & Straub, 2011), SRMR < 0.08 (Hu &
Bentler, 1999).
6. However, researchers must be aware of possible covariance-equivalent models.
7. When used with the default settings, some PLS software does not bootstrap the PLS algorithm, but a mod-
ified version that is designed to produce smaller standard errors. We also estimated our example with this
construct level sign change correction, but the results were no closer to the real distribution.
8. Chin cites Wold (1985b), which uses Wold (1985a) as an example of a small-sample study when discussing
model cross-validation.
References
Aguirre-Urreta, M. I., Marakas, G. M., & Ellis, M. E. (in press). Measurement of composite reliability in
research using partial least squares: Some issues and an alternative approach. DATA BASE for Advances
in Information Systems.
Antonakis, J., Bendahan, S., Jacquart, P., & Lalive, R. (2010). On making causal claims: A review and recom-
mendations. Leadership Quarterly, 21, 1086-1120.
Ashill, N. J., & Jobber, D. (2010). Measuring state, effect, and response uncertainty: Theoretical construct
development and empirical validation. Journal of Management, 36(5), 1278-1308.
Asparouhov, T., & Muthen, B. (2009). Exploratory structural equation modeling. Structural Equation
Modeling: A Multidisciplinary Journal, 16(3), 397-438. doi:10.1080/10705510903008204
Atinc, G., Simmering, M. J., & Kroll, M. J. (2011). Control variable use and reporting in macro and micro man-
agement research. Organizational Research Methods, 15(1), 57-74.
Downloaded from orm.sagepub.com at NATIONAL CHUNG HSING UNIV on April 12, 2014
Ronkko and Evermann 445
Avolio, B. J., Howell, J. M., & Sosik, J. J. (1999). A funny thing happened on the way to the bottom line: Humor
as a moderator of leadership style effects. Academy of Management Journal, 42(2), 219-227. doi:10.2307/
257094
Bass, B. M., Avolio, B. J., Jung, D. I., & Berson, Y. (2003). Predicting unit performance by assessing transfor-
mational and transactional leadership. Journal of Applied Psychology, 88(2), 207-218.
Birkinshaw, J., Hood, N., & Jonsson, S. (1998). Building firm-specific advantages in multinational corpora-
tions: The role of subsidiary initiative. Strategic Management Journal, 19(3), 221-242.
Birkinshaw, J., Morrison, A., & Hulland, J. (1995). Structural and competitive determinants of a global integra-
tion strategy. Strategic Management Journal, 16(8), 637-655.
Bollen, K. A. (1989). Structural equations with latent variables. New York, NY: John Wiley.
Bollen, K. A., Kirby, J. B., Curran, P. J., Paxton, P. M., & Chen, F. (2007). Latent variable models under mis-
specification: Two-stage least squares (2SLS) and maximum likelihood (ML) estimators. Sociological
Methods & Research, 36(1), 48-86. doi:10.1177/0049124107301947
Bollen, K. A., & Stine, R. A. (1992). Bootstrapping goodness-of-fit measures in structural equation models.
Sociological Methods & Research, 21(2), 205-229. doi:10.1177/0049124192021002004
Burt, R. S. (1976). Interpretational confounding of unobserved variables in structural equation models.
Sociological Methods & Research, 5(1), 3-52. doi:10.1177/004912417600500101
Cheung, M.-S., Myers, M. B., & Mentzer, J. T. (2011). The value of relational learning in global buyer-supplier
exchanges: A dyadic perspective and test of the pie-sharing premise. Strategic Management Journal,
32(10), 1061-1082. doi:10.1002/smj.926
Chin, W. W. (1998). The partial least squares approach to structural equation modeling. In G. A. Marcoulides
(Ed.), Modern methods for business research (pp. 295-336). Mahwah, NJ: Lawrence Erlbaum.
Chin, W. W., & Newsted, P. R. (1999). Structural equation modeling analysis with small samples using partial
least squares. Statistical Strategies for Small Sample Research, 2, 307-342.
Cohen, J., Cohen, P., West, S. G., & Aiken, L. S. (2003). Applied multiple regression/correlation analysis for
the behavioral sciences. London, England: Lawrence Erlbaum.
Cool, K., Dierickx, I., & Jemison, D. (1989). Business strategy, market structure and risk-return relationships:
A structural approach. Strategic Management Journal, 10(6), 507-522.
Cording, M., Christmann, P., & King, D. R. (2008). Reducing causal ambiguity in acquisition integration:
Intermediate goals as mediators of integration decisions and acquisition performance. Academy of
Management Journal, 51(4), 744-767.
Crossland, C., & Hambrick, D. C. (2011). Differences in managerial discretion across countries: How nation-
level institutions affect the degree to which CEOs matter. Strategic Management Journal, 32(8), 797-819.
doi:10.1002/smj.913
Delios, A., & Beamish, P. W. (1999). Geographic scope, product diversification and the corporate performance
of Japanese firms. Strategic Management Journal, 20(8), 711-727.
Dijkstra, T. K. (1983). Some comments on maximum likelihood and partial least squares methods. Journal of
Econometrics, 22(1-2), 67-90.
Duxbury, L. E., & Higgins, C. A. (1991). Gender differences in work-family conflict. Journal of Applied
Psychology, 76(1), 60-73. doi:10.1037/0021-9010.76.1.60
Echambadi, R., Campbell, B., & Agarwal, R. (2006). Encouraging best practice in quantitative management
research: An incomplete list of opportunities. Journal of Management Studies, 43(8), 1801-1820. doi:
10.1111/j.1467-6486.2006.00660.x
Edwards, J. R. (2011). The fallacy of formative measurement. Organizational Research Methods, 14(2),
370-388. doi:10.1177/1094428110378369
Evermann, J., & Tate, M. (2010). Testing models or fitting models? Identifying model misspecification in PLS.
In ICIS 2010 proceedings. Retrieved from http://aisel.aisnet.org/icis2010_submissions/21
Fornell, C., & Bookstein, F. L. (1982). Two structural equation models: LISREL and PLS applied to consumer
exit-voice theory. Journal of Marketing Research, 19(4), 440-452.
Downloaded from orm.sagepub.com at NATIONAL CHUNG HSING UNIV on April 12, 2014
446 Organizational Research Methods 16(3)
Fornell, C., & Larcker, D. F. (1981). Evaluating structural equation models with unobservable variables and
measurement error. Journal of Marketing Research, 18(1), 39-50.
Gefen, D., Rigdon, E. E., & Straub, D. W. (2011). An update and extension to SEM guidelines for administra-
tive and social science research. MIS Quarterly, 35(2), iii-xiv.
Goodhue, D. L., Lewis, W., & Thompson, R. (2012). Comparing PLS to regression and LISREL: A response to
Marcoulides, Chin, and Saunders. MIS Quarterly, 36(3), 703-716.
Groth, M., Hennig-Thurau, T., & Walsh, G. (2009). Customer reactions to emotional labor: The roles of
employee acting strategies and customer detection accuracy. Academy of Management Journal, 52(5),
958-974.
Gruber, M., Heinemann, F., Brettel, M., & Hungeling, S. (2010). Configurations of resources and capabilities
and their performance implications: An exploratory study on technology ventures. Strategic Management
Journal, 31(12), 1337-1356.
Hair, J. F., Ringle, C. M., & Sarstedt, M. (2011). PLS-SEM: Indeed a silver bullet. Journal of Marketing Theory
& Practice, 19(2), 139-152. doi:10.2753/MTP1069-6679190202
Hair, J. F., Sarstedt, M., Pieper, T. M., & Ringle, C. M. (2012). The use of partial least squares structural equa-
tion modeling in strategic management research: A review of past practices and recommendations for future
applications. Long Range Planning, 45(5-6), 320-340. doi:10.1016/j.lrp.2012.09.008
Henseler, J., & Chin, W. W. (2010). A comparison of approaches for the analysis of interaction effects
between latent variables using partial least squares path modeling. Structural Equation Modeling: A
Multidisciplinary Journal, 17(1), 82-109. doi:10.1080/10705510903439003
Henseler, J., & Sarstedt, M. (2013). Goodness-of-fit indices for partial least squares path modeling.
Computational Statistics, 28(2), 565-580. doi:10.1007/s00180-012-0317-1
Herzog, W., & Boomsma, A. (2009). Small-sample robust estimators of noncentrality-based and incremental
model fit. Structural Equation Modeling: A Multidisciplinary Journal, 16(1), 1-27. doi:10.1080/
10705510802561279
Howell, J. M., & Avolio, B. J. (1993). Transformational leadership, transactional leadership, locus of control,
and support for innovation: Key predictors of consolidated-business-unit performance. Journal of Applied
Psychology, 78(6), 891-902. doi:10.1037/0021-9010.78.6.891
Howell, J. M., & Hall-Merenda, K. E. (1999). The ties that bind: The impact of leader-member exchange, trans-
formational and transactional leadership, and distance on predicting follower performance. Journal of
Applied Psychology, 84(5), 680-694.
Hu, L., & Bentler, P. M. (1999). Cutoff criteria for fit indexes in covariance structure analysis: Conventional
criteria versus new alternatives. Structural Equation Modeling: A Multidisciplinary Journal, 6(1), 1-55.
Hulland, J. (1999). Use of partial least squares (PLS) in strategic management research: A review of four recent
studies. Strategic Management Journal, 20(2), 195-204.
Johansson, J. K., & Yip, G. S. (1994). Exploiting globalization potential: U.S. and Japanese strategies. Strategic
Management Journal, 15(8), 579-601.
Joreskog, K. G., & Wold, H. (1982). The ML and PLS techniques for modeling with latent variables. In K. G.
Joreskog & H. Wold (Eds.), Systems under indirect observation: Causality, structure, prediction (pp.
263-270). Amsterdam, the Netherlands: North-Holland.
Lai, K., & Kelley, K. (2011). Accuracy in parameter estimation for targeted effects in structural equation
modeling: Sample size planning for narrow confidence intervals. Psychological Methods, 16(2), 127-148.
doi:10.1037/a0021764
Landis, R. S., Beal, D. J., & Tesluk, P. E. (2000). A comparison of approaches to forming composite
measures in structural equation models. Organizational Research Methods, 3(2), 186-207. doi:10.1177/
109442810032003
Lehmann, E., & Casella, G. (1998). Theory of point estimation. New York, NY: Springer.
Little, T. D., Cunningham, W. A., Shahar, G., & Widaman, K. F. (2002). To parcel or not to parcel: Exploring
the question, weighing the merits. Structural Equation Modeling, 9(2), 151-173.
Downloaded from orm.sagepub.com at NATIONAL CHUNG HSING UNIV on April 12, 2014
Ronkko and Evermann 447
Lohmoller, J. B. (1989). Latent variable path modeling with partial least squares. Heidelberg, Germany:
Physica-Verlag.
McDonald, R. P. (1996). Path analysis with composite variables. Multivariate Behavioral Research, 31(2),
239-270.
Meznar, M. B., & Nigh, D. (1995). Buffer or bridge? Environmental and organizational determinants of public
affairs activities in American firms. Academy of Management Journal, 38, 975-996.
Mulaik, S. A. (1985). Exploratory statistics and empiricism. Philosophy of Science, 52(3), 410-430.
Nickerson, R. S. (2000). Null hypothesis significance testing: A review of an old and continuing controversy.
Psychological Methods, 5(2), 241-301. doi:10.1037//1082-989X.5.2.241
Nunnally, J. (1978). Psychometric theory. New York, NY: McGraw-Hill.
Olk, P., & Young, C. (1997). Why members stay in or leave an R&D consortium: Performance and conditions
of membership as determinants of continuity. Strategic Management Journal, 18(11), 855-877.
Peng, D. X., & Lai, F. (2012). Using partial least squares in operations management research: A practical guideline and
summary of past research. Journal of Operations Management, 30(6), 467-480. doi:10.1016/j.jom.2012.06.002
Raykov, T. (1997). Estimation of composite reliability for congeneric measures. Applied Psychological
Measurement, 21(2), 173-184. doi:10.1177/01466216970212006
Reinartz, W. J., Haenlein, M., & Henseler, J. (2009). An empirical comparison of the efficacy of
covariance-based and variance-based SEM. International Journal of Research in Marketing, 26(4),
332-344. doi:10.1016/j.ijresmar.2009.08.001
Robins, J. A., Tallman, S., & Fladmoe-Lindquist, K. (2002). Autonomy and dependence of international coop-
erative ventures: An exploration of the strategic performance of U.S. ventures in Mexico. Strategic
Management Journal, 23(10), 881-901. doi:10.1002/smj.260
Rosseel, Y. (2012). Lavaan: An R package for structural equation modeling. Journal of Statistical Software,
48(2), 1-36.
Sanchez, G., & Trinchera, L. (2012). plspm version 0.2-2 [Computer software]. Retrieved from http://cran.r-
project.org/web/packages/plspm/index.html
Sarkar, M. B., Echambadi, R. A. J., & Harrison, J. S. (2001). Alliance entrepreneurship and firm market
performance. Strategic Management Journal, 22(6/7), 701-711. doi:10.1002/smj.179
Shamir, B., Zakay, E., Breinin, E., & Popper, M. (1998). Correlates of charismatic leader behavior in military
units: Subordinates attitudes, unit characteristics, and superiors appraisals of leader performance. Academy
of Management Journal, 41, 387-409.
Shea, C. M., & Howell, J. M. (2000). Efficacy-performance spirals: An empirical test. Journal of Management,
26(4), 791-812. doi:10.1177/014920630002600409
Sosik, J. J., Avolio, B. J., & Kahai, S. S. (1997). Effects of leadership style and anonymity on group potency and
effectiveness in a group decision support system environment. Journal of Applied Psychology, 82(1),
89-103.
Sosik, J. J., Kahai, S. S., & Piovoso, M. J. (2009). Silver bullet or voodoo statistics? A primer for using the
partial least squares data analytic technique in group and organization research. Group & Organization
Management, 34(1), 5-36. doi:10.1177/1059601108329198
Tiwana, A. (2008). Do bridging ties complement strong ties? An empirical examination of alliance ambidex-
terity. Strategic Management Journal, 29(3), 251-272.
Tsang, E. W. K. (2002). Acquiring knowledge by foreign partners from international joint ventures in a transi-
tion economy: Learning-by-doing and learning myopia. Strategic Management Journal, 23(9), 835-854.
doi:10.1002/smj.251
Vandenberg, R. J. (2006). Introduction: Statistical and methodological myths and urban legends.
Organizational Research Methods, 9(2), 194-201. doi:10.1177/1094428105285506
Vinzi, V. E., Chin, W. W., Henseler, J., & Wang, H. (2010). Editorial: Perspectives on partial least squares. In
V. E. Vinzi, W. W. Chin, J. Henseler, & H. Wang (Eds.), Handbook of partial least squares (pp. 1-22).
Berlin, Germany: Springer.
Downloaded from orm.sagepub.com at NATIONAL CHUNG HSING UNIV on April 12, 2014
448 Organizational Research Methods 16(3)
Wang, T., & Bansal, P. (2012). Social responsibility in new ventures: Profiting from a long-term orientation.
Strategic Management Journal, 33, 1135-1153. doi:10.1002/smj.1962
Wold, H. (1985a). Factors influencing the outcome of economic sanctions. Trabajos de Estadistica y de
Investigacion Operativa, 36(3), 325-338. doi:10.1007/BF02888567
Wold, H. (1985b). Systems analysis by partial least squares. In P. Nijkamp, L. Leitner, & N. Wrigley (Eds.),
Measuring the unmeasurable (pp. 221-252). Dordrecht, Germany: Marinus Nijhoff.
Wood, M. (2005). Bootstrapped confidence intervals as an approach to statistical inference. Organizational
Research Methods, 8(4), 454-470. doi:10.1177/1094428105280059
Zimmerman, D. W., & Williams, R. H. (1977). The theory of test validity and correlated errors of measurement.
Journal of Mathematical Psychology, 16(2), 135-152.
Zott, C., & Amit, R. (2008). The fit between product market strategy and business model: Implications for firm
performance. Strategic Management Journal, 29(1), 1-26.
Author Biographies
Mikko Ronkko is a doctoral candidate at Aalto University, School of Science. His research interests are in
statistics and research methods, with a focus on structural equation modeling.
Joerg Evermann is an associate professor of information systems. His research interests are in statistics and
research methods, with a focus on structural equation modeling.
Downloaded from orm.sagepub.com at NATIONAL CHUNG HSING UNIV on April 12, 2014