SSRN Id22337955421
SSRN Id22337955421
com/locate/lrp
Editorial
Partial Least Squares Structural Equation
Modeling: Rigorous Applications, Better
Results and Higher Acceptance
Wold’s (1982, 1974) and Lohm€ oller’s (1989) partial least squares structural equation modeling
(PLS-SEM) approach has enjoyed increasing popularity as a key multivariate analysis method in
various research disciplines such as accounting (Lee et al., 2011), international marketing
(Henseler et al., 2009), management information systems (Ringle et al., 2012), marketing (Hair
et al., 2012c) and operations management (Peng and Lai, 2012). Similarly, strategic management
research has recognized PLS-SEM’s flexibility regarding the handling of various modeling problems
(e.g., Hulland, 1999). However, according to Hair et al.’s. (2012b) recent review of PLS-SEM use
in the strategic management field, the method’s dissemination is not as widespread as in other
fields.
This lower dissemination suggests that PLS-SEM use in strategic management presents numerous
application opportunities. More often than not, researchers do not pay attention to data quality
requirements (e.g., with regard to sample size or treatment of missing values) and disregard impor-
tant evaluation criteria (especially when assessing formative measurement models), sometimes even
misapplying them (e.g., by using reflective measurement model assessment procedures on forma-
tive measurement models). Most studies do not provide sufficient information to enable readers to
replicate the results and fully assess a study’s quality. Consequently, Hair et al. (2012c; p. 430) con-
clusion in the field of marketing also seems to hold for the strategic management discipline: “While
offering many beneficial properties, PLS-SEM’s ‘soft assumptions’ should not be taken as carte
blanche to disregard standard psychometric assessment techniques.” Furthermore, researchers do
not make use of the entire toolbox that PLS-SEM offers. Methodological advances (Esposito
Vinzi et al., 2010; Hair et al., 2013) provide researchers much more flexibility in modeling relation-
ships and thus allow for a more nuanced testing of theoretical concepts. Amongst others, these ad-
vances include confirmatory tetrad analysis (CTA-PLS) to empirically assess the measurement
model type (i.e., formative or reflective; Gudergan et al., 2008), importance-performance matrix
analysis (IPMA) of PLS-SEM results (e.g., H€ ock et al., 2010; Rigdon et al., 2011; V€olckner et al.,
2010), approaches to assess hierarchical component models (e.g., Becker et al., 2012; Ringle
et al., 2012; Wilson, 2010), PLS-SEM-specific data segmentation techniques (e.g., Rigdon et al.,
2010; Sarstedt, 2008), analysis of interaction effects (e.g., Henseler and Chin, 2010; Henseler and
Fassott, 2010) and other nonlinear effects (Dijkstra and Henseler, 2011; Henseler et al., 2012a;
Rigdon et al., 2010) or multigroup analysis procedures (e.g., Chin and Dibbern, 2010; Sarstedt
et al., 2011b).
A summary of the main findings of recent reviews of PLS-SEM applications across a variety of
disciplines suggests that the following issues are prevalent and require the particular attention of
authors, reviewers, and editors to improve the quality of PLS-SEM studies:
*
The authors thank J€
org Henseler (Radboud University Nijmegen) for his valuable comments and suggestions.
Electronic
Electroniccopy
copyavailable
availableat:
at:https://ssrn.com/abstract=2233795
http://ssrn.com/abstract=2233795
Goal of the analysis
Some researchers claim they use PLS-SEM for predictive purposes. However, this aspect often is
not reflected in the analysis and interpretation of results. For example, researchers who claim to
use PLS-SEM for their study’s predictive purpose frequently accept relatively low levels of R2
values and/or do not analyze criteria such as the predictive relevance Q2 and related measures
(Hair et al., 2013; Henseler et al., 2012b).
PLS-SEM is advantageous when used with small sample sizes (e.g., in terms of the robustness of
estimations and statistical power; Reinartz et al., 2009). However, some researchers abuse this
advantage by relying on extremely small samples relative to the underlying population. All else
being equal, the more heterogeneous the population in a structure is the more observations are
needed to reach an acceptable sampling error level. Ignoring the fundamentals of sampling the-
ory yields meaningless results no matter which method is applied. PLS-SEM has an erroneous
reputation for offering special sampling capabilities that no other multivariate analysis tool has.
Like any other statistical technique, inference statistics based on PLS-SEM require representa-
tive samples. Researchers are therefore well-advised to use sampling techniques carefully (which
also applies to all other studies using different analysis techniques) and carefully consider the
statistical power of their analyses (see Hair et al., 2013 for an overview).
Similarly, researcher frequently claim they use PLS-SEM because their data is non-normal. Ad-
ditional descriptive statistics (e.g., skewness and kurtosis) to substantiate this argument and to
inform the reader about the data are routinely omitted, even though they have important im-
plications for the analyses (e.g., with regard to the use of bootstrapping procedures; Hair et al.,
2013). Statements about missing values are also rare, as hardly any studies provide information
on the magnitude of the missing data (e.g., percentages per variable and case) and the missing
value treatment option used (e.g., mean value replacement or case-wise deletion). Providing in-
formation on missing values and potential non-response bias is crucial to maximize confidence
in the results. Furthermore, treatment of missing values has vast implications for PLS-SEM use.
For example, even if a relatively small number of missing values has been replaced by the mean
value of the remaining (valid) observations, methods for analyzing unobserved heterogeneity
(e.g., by using FIMIX-PLS; Hahn et al., 2002; Rigdon et al., 2011; Ringle et al., 2010a,b;
Sarstedt et al., 2011a; Sarstedt and Ringle, 2010) will yield highly biased results. Researchers
need to fully report data characteristics and treat missing values carefully in their analyses.
PLS-SEM software applications, such as SmartPLS (Ringle et al., 2005), provide results for all
types of variables, regardless of whether they have metric, quasi-metric, ordinal, or categorical
scales (e.g., binary coded). In its basic form, however, PLS-SEM requires metric or quasi-metric
data. The use of other scale levels changes the interpretation of results or violates the method’s
fundamental requirements.1 For example, one should not use dummy variables in reflective
measurement models. Moreover, the use of dummy variables in formative measurement models
requires an interpretation similar to that for regression analyses with dummy variables (Hair
et al., 2010). Thus, researchers using PLS-SEM in such data constellations should be acquainted
1
Basically, PLS-SEM builds on the ordinary least squares method, which requires the dependent variable to be continuous. If it
is not, this requirement is violated when estimating a PLS path model, which has significant consequences for the analysis.
2 Editorial
Electronic
Electroniccopy
copyavailable
availableat:
at:https://ssrn.com/abstract=2233795
http://ssrn.com/abstract=2233795
with the basic principles of the PLS-SEM algorithm to understand which model set-ups allow
for the use of categorical variables.
Researchers often use PLS-SEM because this technique facilitates the inclusion of formative
measurement models, which have recently attracted considerable attention across various dis-
ciplines (e.g., Diamantopoulos and Riefler, 2011; Diamantopoulos et al., 2008; Diamantopoulos
and Winklhofer, 2001; Gudergan et al., 2008; Jarvis et al., 2003; MacKenzie et al., 2005;
MacKenzie et al., 2011). However, formative measurement models should be evaluated cau-
tiously, for instance, by following the recommendations given by Hair et al. (2013) in the con-
text of PLS-SEM. Like reflective measurement models, they require a thorough evaluation, for
example, by assessing the indicator weights, the loadings, or by carrying out redundancy anal-
yses.2 If formative measurement models are not carefully evaluated the value of the entire anal-
ysis is in doubt. Finally, the CTA-PLS analysis (Gudergan et al., 2008) allows distinguishing
empirically a formative measurement model specification from a reflective one. All these aspects
are usually not taken into account when researchers include formative measurement models in
their PLS-SEM study.
Control variables
Control variables are often included in PLS path models, accounting for some of the target
construct’s variation. Regardless of whether these control variables are significant or not,
the results for control variables are usually not further interpreted. When the effect of control
variables are significant, the researcher should use this finding especially carefully when draw-
ing conclusions or initiating additional analyses (e.g., PLS-SEM multigroup analyses; Sarstedt
et al., 2011b), which occurs routinely with regression-based models (e.g., Raithel et al., 2012).
Lastly, researchers usually neglect the fact that the inclusion of control variables increases
model complexity and, thus, may also increase the required sample size for estimating the
PLS path model.
Moderator analysis
Two crucial issues often lead to problems in moderator analysis. First, the moderators are
included with their interaction terms (i.e., the multiplication of indicators or constructs) in
the PLS path model and their simple effects are mistakenly interpreted as main effects
(Jaccard and Turrisi, 2003). The character of these results differs completely, however, and
requires particular attention when analyzing PLS-SEM results (Hair et al., 2013; Henseler
and Chin, 2010; Henseler and Fassott, 2010).3 One may first estimate and evaluate the
main effects in the PLS path model and, in a subsequent moderator analysis, include the prod-
uct term and its interaction effect to avoid the common mistake of confounding main and
simple effects. As an additional remedy, researchers should consider using orthogonalization
(Henseler and Chin, 2010). Second, when there is theoretical support for multiple moderators,
2
Note that standard assessment procedures for formative measurement models disregard the loadings. Recent research however
(Cenfetelli and Bassellier, 2009; Hair et al., 2013) suggests that in situations in which the weights are not significant, researchers
should also consider the indicators’ absolute contribution to the construct (i.e., the loadings).
3
A PLS path model without moderation includes only main effects between latent variables in the structural model. The main
effects model becomes a moderator model after including a product term and its interaction (or moderating) effect. In a moderator
model, main effects change into simple (or single) effects (Henseler and Fassott, 2010). This distinction is important and must be
properly accounted for when interpreting PLS-SEM results. Whereas a main effect quantifies the change in the level of the depen-
dent variable when the considered independent variable is increased by one unit and all other independent variables remain con-
stant (ceteris paribus), a simple effect quantifies the change in the level of the dependent variable when the independent variable is
increased by one unit, the interacting variable has a value of zero, and all other independent variables remain constant. Henseler
and Fassott (2010) and Hair et al. (2013) provide detailed explanations and examples on testing moderating effects in PLS-SEM.
Non-linear models
The ever-increasing complexity of theories and model constellations has increased the promi-
nence of the modeling of non-linear effects (Cortina, 1993). While nonlinear effects can be eas-
ily included in a PLS path model (Dijkstra and Henseler, 2011; Henseler et al., 2012a; Rigdon
et al., 2010), their actual use is a matter of concern. First, researchers mostly do not provide
a precise a priori reasoning for the shape of the non-linear effect. Thereby, they do not substan-
tiate their hypotheses. Merely searching for a nonlinear effect entails capitalizing on chance re-
sults. The more you search in many different directions, the higher the probability you will find
something. However, it is very unlikely that any replication of such an ex post facto analysis
would yield exactly the same effect (Cliff, 1983). Consequently, the corresponding outcomes
are not generalizable. Second, even though nonlinear effects are frequently significant, the ques-
tion remains regarding their added value. A critical look at the outcomes often indicates signif-
icant but relatively small sizes of the quadratic effect, entailing almost no changes in the R2
values. In such cases, a linear and, thus simpler, model (i.e., more parsimonious) provides al-
most the same results. In the tradition of Ockham’s razor, researchers usually opt for the more
parsimonious approach if admissible. Third, the interpretation of nonlinear effects requires the
researcher to characterize the different fractions of a function. Consider, for example, a u-
shaped relationship depending on the size of firms: the fraction with a strong negative slope
for smaller firms, the fraction with a slope of approximately zero for mid-sized firms, and
the fraction with a strong positive slope for large firms. The different fractions may be explained
by the source of the nonlinear effect itself. For this purpose, a PLS-SEM multigroup analysis
(Chin and Dibbern, 2010; Sarstedt et al., 2011b) can be used to approximate the different frac-
tions of a non-linear relationship through group-specific linear regressions with different slope
sizes and/or signs. Again, this approach allows the researcher to switch from the more complex
non-linear design to the more parsimonious linear model while including d as an additional
explanation d the grouping of the data.
Mediator analysis
Most structural models are subject to mediation effects (e.g., Hair et al., 2013; Helm et al.,
2010), which many researchers overlook in their PLS-SEM analyses. In the extreme, they do
not examine and interpret the result of a full mediation but simply state that a relationship be-
tween two latent variables is not significant. Hence, they erroneously conclude that, in the struc-
tural model, the relationship between the two latent variables is nil. Researchers should
routinely report the total effects (i.e., the sum of direct and indirect effects between two con-
structs; Albers, 2010). This not only allows a more complete picture of the mediating constructs’
role, but also provides practitioners with actionable results regarding cause-effect relationships.
Moreover, formalized mediation analysis by means of bootstrapping analysis (Hair et al., 2013)
is especially valuable when corresponding hypotheses have been formulated (e.g., Sattler et al.,
2010). A final note of caution refers to the quality of measurement models in the context of
mediation. When using variance-based structural equation modeling techniques such as PLS-
4 Editorial
Hierarchical component models are relatively easy to conduct in PLS-SEM by using Wold’s
(1982) repeated indicator approach. Recent explications provide additional insights into
how hierarchical component models should be incorporated into PLS path models (Becker
et al., 2012; Hair et al., 2013; Ringle et al., 2012; Wetzels et al., 2009). Unfortunately, users of-
ten misapply higher order models both conceptually and technically. Instead of presenting
proper explications that justify the type of higher order model used (e.g., the reflective-
reflective type or the reflective-formative type), researchers routinely use hierarchical compo-
nent models to summarize information in a second, third or even higher dimension of abstrac-
tion with the plausible objective of reducing the number of relationships in the PLS path
model. The resulting higher order models however are often difficult to defend from a theoret-
ical point of view.
Despite its ease of application, the repeated indicator approach may entail several pitfalls. First,
the number of indicators per lower order component should be balanced (Becker et al., 2012).
Otherwise the estimated relationships between the higher and the lower order components may
be biased. Second, researchers usually do not evaluate the higher order constructs, although the
same evaluation criteria (and their critical values) used for the lower order components apply.
Hence, information about relevant evaluation criteria outcomes is important and should be
provided. Third, some researchers include relationships from other latent variables in the struc-
tural model, which are not part of the hierarchical component model, to formative higher order
constructs. These relationships always have a value of approximately zero when the indicator
reuse technique is applied to determine the higher order construct in PLS-SEM, because the
formative lower order components already explain all of the former’s variance. Hence, the con-
clusion that other constructs in the structural model do not explain any variation of the
reflective-formative type or the formative-formative type higher order construct would be false
and misleading, as this is a technical outcome of the repeated indicator approach. In such sit-
uations, a two-stage approach (Ringle et al., 2012) should be used, which allows for handling
this technical limitation of the repeated indicator approach.
The results evaluation and reporting are often incomplete. In some instances, the path coefficients
and their significance are not reported, or the R2 values of the endogenous latent variables are missing.
Even if a study shows all the relevant criteria as suggested, for example, by Hair et al. (2013), the re-
searchers often report the results without critical reflection or further interpretation. A fundamental
issue relates to the use of the Goodness-of-Fit (GoF) index proposed by Tenenhaus et al. (2004, 2005)
as a means to validate a PLS path model globally. Henseler and Sarstedt (2013) challenged the use-
fulness of the GoF index conceptually and empirically. For instance, in a simulation study, the authors
show that the GoF index cannot separate valid models from invalid ones. Moreover, the GoF index is
not applicable to formative measurement models and does not penalize overparameterization efforts.
Furthermore, the software used (for a PLS software comparison see Temme et al., 2010) is usu-
ally not reported, although this information provides important details regarding the default
values used in the running of the PLS-SEM algorithm and supplementary analyses. When au-
thors do provide this important information, they usually fail to correctly cite the two major
software applications SmartPLS (Ringle et al., 2005) and PLS-Graph (Chin, 2003) as required
in these software’s license agreements (i.e., more than 50 percent of PLS-SEM applications
are subject to this flaw).
1. Data characteristics
Assess the required sample size for estimating the established PLS path model to ensure
a sufficient level of statistical power,
Provide a comprehensive description of the sample (e.g., report the number of observations
and amount of missing values but also mean values, variances and other descriptive statistics),
Characterize the distribution of the variables (e.g., report skewness and kurtosis of data),
Use a holdout sample (e.g., conduct the analysis with 70% of the original sample),
Offer all relevant information to facilitate replication of your analysis (e.g., share the cor-
relation/covariance matrix or raw data in an online appendix),
Explain in detail the scales of variables; treat variables with scales other than metric or
quasi-metric with particular care in PLS-SEM (e.g., do not use categorical variables in en-
dogenous constructs; interpret categorical variables in exogenous constructs carefully).
2. Model characteristics
Fully describe the structural model (i.e., the latent variables and their relationships) using
a graphical (instead of a formal/mathematical) representation of the PLS path model, which
makes it much easier for the reader to quickly grasp some of the key research contents.
Characterize the measurement models (formative vs. reflective) of the latent variables (e.g., em-
pirically substantiate the selected measurement model by using CTA-PLS) and include a com-
plete list of the indicators employed in the measurement models (e.g., in the appendix).
.the starting values of the weights for the initial approximation of the latent variable
scores (e.g., use a uniform value of 1 as an initial value for each of the outer weights),
.the inner weighting scheme to determine the latent variable scores,
.the stop criterion (e.g., the sum of the measurement model weights’ absolute changes be-
tween two iterations <0.00001),
.the actual number of iterations (e.g., 300),
.the software used (e.g., report the software, including the version to indicate the default
settings).
4. Report the exact algorithm settings used when running supplementary procedures (e.g., bootstrap-
ping, blindfolding, CTA-PLS, analysis of moderating effects, multi-group analysis, FIMIX-PLS) in detail.
4
Besides the guidelines given in this editorial, one finds various recommendations on the use of the PLS-SEM method and the
evaluation of results in prior literature; examples include: Chin (1998, 2010); Falk and Miller (1992); G€
otz et al. (2010); Haenlein
and Kaplan (2004); Hair et al. (2011); Hair et al. (2012); Henseler et al. (2009); Lohm€ oller (1989); Roldan and Sanchez-Franco
(2012); Sosik et al. (2009); Tenenhaus et al. (2005).
6 Editorial
These guidelines, as well as new textbooks on the PLS-SEM method (e.g., Hair et al., 2013) pro-
vide researchers, editors, and reviewers with the knowledge they need to ensure the rigor of pub-
lished research in academic journals. By following these recommendations, the quality of studies
employing PLS-SEM should be enhanced and crucial mistakes avoided. As in any empirical re-
search, the goal in PLS-SEM is to progress towards the highest possible level of transparency
that allows the replication of published studies. Future developments towards this goal will substan-
tially improve the way in which research is conducted, as well as the quality of published articles.
This second Long Range Planning (http://www.journals.elsevier.com/long-range-planning/) special
issue on PLS-SEM in strategic management research and practice seeks to further progress towards this
goal. The journal received 41 articles for its special issue on PLS-SEM, twelve of which completed a thor-
ough review process successfully. Based on the number of high quality manuscripts, a decision was
made to split the special issue. In the first Long Range Planning special issue on PLS-SEM in strategic
management (Hair et al., 2012a; Robins, 2012), the focus was on methodological developments and
their application (Becker et al., 2012; Furrer et al., 2012; Gudergan et al., 2012; Hair et al., 2012a,b,c;
Money et al., 2012; Rigdon, 2012). This second special issue provides a forum for topical issues that
demonstrate the usefulness of PLS-SEM by piloting applications of this method in the field of strategic
management with strong implications for strategic research and practice. As such, the special issue tar-
gets two audiences: academics involved in the fields of strategy and management, and practitioners such
as consultants. The six articles in this issue are summarized in the following paragraphs.
In their paper “Crossing Borders and Industry Sectors: Behavioral Governance in Strategic Alli-
ances and Product Innovation for Competitive Advantage,” Yong Kyu Lew and Rudolf R. Sinkovics
investigate how international technology alliances (ITAs) between software and hardware firms in
8 Editorial
References
Albers, S., 2010. PLS and success factor studies in marketing, In: Esposito Vinzi, V., Chin, W.W., Henseler, J.,
Wang, H. (Eds.), Handbook of Partial Least Squares: Concepts, Methods and Applications (Springer Handbooks
of Computational Statistics Series, vol. II). Springer, Heidelberg, Dordrecht, London, New York, pp. 409e425.
Becker, J.-M., Klein, K., Wetzels, M., 2012. Hierarchical latent variable models in PLS-SEM: guidelines for
using reflective-formative type models, Long Range Planning 45 (5/6), 359e394.
Cenfetelli, R.T., Bassellier, G., 2009. Interpretation of formative measurement in information systems research,
MIS Quarterly 33 (4), 689e708.
Chin, W.W., 1998. The partial least squares approach to structural equation modeling, In: Marcoulides, G.A.
(Ed.), Modern Methods for Business Research. Erlbaum, Mahwah, pp. 295e358.
Chin, W.W., 2003. PLS Graph 3.0. Soft Modeling Inc, Houston.
Chin, W.W., 2010. How to write up and report PLS analyses, In: Esposito Vinzi, V., Chin, W.W., Henseler, J.,
Wang, H. (Eds.), Handbook of Partial Least Squares: Concepts, Methods and Applications (Springer
Handbooks of Computational Statistics Series, vol. II). Springer, Heidelberg, Dordrecht, London, New
York, pp. 655e690.
Chin, W.W., Dibbern, J., 2010. A permutation based procedure for multi-group PLS analysis: results of tests
of differences on simulated data and a cross cultural analysis of the sourcing of information system services
between Germany and the USA, In: Esposito Vinzi, V., Chin, W.W., Henseler, J., Wang, H. (Eds.), Hand-
book of Partial Least Squares: Concepts, Methods and Applications (Springer Handbooks of Computa-
tional Statistics Series, vol. II). Springer, Heidelberg, Dordrecht, London, New York, pp. 171e193.
Cliff, N., 1983. Some cautions concerning the application of causal modeling methods, Multivariate Behav-
ioral Research 18 (1), 115e126.
Cortina, J.M., 1993. Interaction, nonlinearity, and multicollinearity: implications for multiple regression,
Journal of Management 19 (4), 915e922.
Diamantopoulos, A., Riefler, P., 2011. Using formative measures in international marketing models: a caution-
ary tale using consumer animosity as an example, Advances in International Marketing 10 (22), 11e30.
Diamantopoulos, A., Riefler, P., Roth, K.P., 2008. Advancing formative measurement models, Journal of Busi-
ness Research 61 (12), 1203e1218.
10 Editorial
Biographies
Joseph F. Hair, Jr., Kennesaw State University
Christian M. Ringle, Hamburg University of Technology (TUHH)
Marko Sarstedt, Otto-von-Guericke-University Magdeburg
12 Editorial