Meta - Note 2
Meta - Note 2
1
At least use two different databases (DeSimone et al., 2020), to prevent biased results due to the scope or journal coverage of
one database.
Gusenbauer & Haddaway (2020) point out that Google Scholar is not appropriate as a primary search engine due to a lack of
reproducibility of search results.
o Report the order of application of exclusion rules and the number of studies remaining after
each step
o
Level of analysis
o Explicitly specify the level(s) (i.e., individual? team? organization? and/or industry?) of interest
for the phenomena under investigation and how your search strategy yield appropriate data.
o Research has demonstrated the fallacy of generalizing across levels (Glick, 1985; Rousseau,
1985).
study inclusion criteria and sample composition
o principle: inclusion of all available studies (Steel et al., 2021)2
o the screening process should be conducted stepwise:
1.remove duplicate citations from different databases
2.abstrat screening to exclude clearly unsuitable studies
3.full-text screening
o A graphical tool to document the sample selection process: PRISMA (Moher et al., 2009; Page
et al., 2021)
o exclusion criteria:
use unsuitable variable measurement
not report usable effect size
multiple studies by the same authors using similar datasets (if they don’t differ
sufficiently the sample characteristics or variables used, only one of these studies should
be included, see (Wood, 2008)for a detection heuristic)
3
The workbooks can be downloaded from: https://www.erim.eur.nl/research-support/meta-essentials/.
R: metafor package4 (Viechtbauer, 2010)
SPSS, SAS: existing macros5
Step 7. Analysis
Outlier analysis
o Aim: ensure the robustness of the meta-analytical findings.
o should explicitly report and justify how outliers were identified and addressed
o Ways to identify outliers (Aguinis et al., 2013; Viechtbauer & Cheung, 2010)
sample-adjusted meta-analytic deviancy statistic (Beal, Corey, & Dunlap, 2002), and
analogs to ordinary regression diagnostics (Viechtbauer & Cheung, 2010).
o Ways to deal with outliers
the leave- one-out analyses (Viechtbauer, 2010), or
removing outliers in groups (Schmidt & Hunter, 2015)
o Whether to keep or omit the outliers depends on:
Outliers can be a valid representation: Representing a different population,
measure, construct, etc.
Provide the basis to infer potential moderators (Aguinis et al., 2013)
Outliers can indicate invalid research (e.g., construct overlap), invalid measures,
or typing errors.
Advisable step: compare the results both with and without outliers and base the
decision on whether to exclude outlier observations.
Consider both size and leverage (Viechtbauer & Cheung, 2010) of the
outlier
Publication bias
o Methods to examine publication bias:
Funnel plot (Stanley & Doucouliagos 2010): can help detect publication bias and outlier;
Statistical procedures:
subgroup comparisons between published and unpublished studies
rank correlation test (Begg & Mazumdar, 1994)
cumulative meta-analysis (Borenstein et al. 2009)
the trim and fill method (Duval and Tweedie 2000a, b)
Egger et al.’s (1997) regression test
fail- safe N (Rosenthal 1979)
selection models (Hedges & Vevea 2005; Vevea & Woods 2005).
Recommend “triangulation approach”: use multiple tests to examine it (Harrison et al.,
2017; Kepes et al., 2012)
Availability bias
o i.e., a subset of effect size is not accessible for reasons other than publication bias (e.g.,
studies published in other languages)
o There are several different approaches to detecting availability bias (Rothstein et al., 2005),
and the methods continue to evolve (Banks et al., 2012; O’Boyle et al., 2014)
drift analysis (Banks et al., 2012)
funnel plots (Peters et al., 2008)
trim-and-fill analysis (Duval & Tweedie, 2000)
o should report and justify the methods used for any analyses conducted; consider evidence
from multiple indices of availability bias (DeSimone et al., 2021)
Psychometric adjustments (corrections) for statistical artifacts
o i.e., effect size estimates may be adjusted to account for the influence of various artifacts
o Common psychometric adjustments:
Sampling error (which is addressed by weighting primary study effect sizes)
attenuation due to measurement error (i.e., imperfect measurement reliability)
report information about primary study reliability estimates and type(s) of
reliability used
reliability estimates (e.g., alpha, retest, interrater) are not interchangeable
(DeSimone et al., 2021)
range restriction
?
o should report information about each adjustment:
which adjustments were used (e.g., predictor reliability, criterion reliability, range
restriction),
the order in which adjustments were made,
whether the meta-analysts considered potential dependencies between artifacts (Ko ̈hler
et al., 2015),
whether adjustments were applied at the local level (i.e., to individual primary studies;
see Viswesvaran et al., 1996) or the global level (i.e., to the summary effect size; see
Oswald & McCloy, 2003).
o Should report both observed (unadjusted) and adjusted effect sizes and variance estimates
Moderator analysis
Sensitivity analysis
o it can be applied to decisions (e.g., the treatment of missing data, the identification of outliers)
made at any stage of the meta-analysis (e.g., coding, data, analysis)
o authors can conduct sensitivity analysis and report what the results would look like if different
decisions had been made.
o Open-access tools are available to facilitate the conduct of sensitivity analyses (see Field et al.,
2020)
Meta-analytic model6
o Random-effects (RE) model
Assume that the variance in the distribution of observed effects is attributed to
1) within-study variance (i.e., sampling error), and
2) between-study variance (i.e., differences of true effect size in the
superpopulation)
CI about mean effect size is wider than using CE model
Always the more appropriate choice in organizational science research (Aguinis et al.,
2011)
Two assumptions must be satisfied (could seen as limitations):
random-studies assumption: The observed effect sizes have been randomly
sampled from a superpopulation of ture effect sizes
superpopulation7 normality assumption: The superpopulation of ture effect sizes is
normally distributed
o requre at least 20 primary-level studies (Field, 2015)
o Fixed-effect or common-effect (CE) model
Assume that the variance in the distribution of observed effects is attributed solely to
1) within-study variance (i.e., sampling error)
6
The decision of model is based on 1) the goal of meta-analysis and 2) the nature of the primary-level studies included in the
review. Some meta-analysis assessed Q (i.e., variance in true effect sizes). but Q’s significance level should not be used as a
decision-making tool for choosing among RE, CE, and VC models (Aguinis et al., 2011).
7
Superpopulation:
Appropriate to use when each of the primary-level studies included in a meta-analysis is
functionally identical8
This condition rarely exists in organizational research, could occur in biological,
medical, and health sciences
Advantages
Can be used with as few as 2 studies (Borenstein et al., 2009, p.363)
o Varying coefficient (VC) model
Advantages
Do not assume equality of effect sizes across studies
Do not rely on random-studies and superpopulation normality assumptions
Availability of CI for assessing the magnitude of interaction effect
Can be used with as few as 2 studies (Borenstein et al., 2009, p.363)
Limitations
Provide inferences only to a subset of the superpopulation
in organization science, RE and VC models are almost always preferred to the CE model;
RE methods are preferred over the VC methods if the random-studies and
superpopulation normality assumptions can be met.
Step 8. Report results
Checklist for quantitative meta-analysis reporting standards (DeSimone et al., 2021; Page et al.,
2021)
o Table reporting results should include:
Mean effect sizes, standard errors, confidence or credibility intervals 9 around summary
effect sizes, the number of observations, study samples included 10, heterogeneity
measures
8
Identitical = include samples of participants from exactly use the sample population, the same researchers, research design,
and measures for IV and DV
9
DeSimone et al. (202) recommend report prediction interval as well.
10
If the meta-analytic sample is rather small, a forest plot should be provided (Hansen et al., 2022).
Essential elements include (a) the citation, (b) the computed effect size, and (c)
information necessary to calculate the sampling variance of the effect size (e.g., sample
size and related information, such as sample size per group for Cohen’s d or Hedges’s g)
Including heterogeneous types of effect sizes and statistics (e.g., p values, t tests, odds
ratios) requires additional assumptions and computational procedures
When effect size transformations are used, reviewers can suggest that authors report the
original effect size (and type) for each primary study, any transformations used, and any
differences between studies that analyzed the focal relationships in different ways
present this information in a table, appendix, or online supplement to ensure it is
accessible to readers
o Authors can also graphically display effect size distributions (DeSimone et al., 2021) using
formats such as a funnel plot or forest plot (Anzures-Cabrera & Higgins, 2010; Kepes et al.,
2012)
Forest plots can depict a great deal of information about an effect size distribution,
including summary effect size magnitude, heterogeneity, and moderation.
o Results displayed in the tables and figures must be explained verbally in the results and
discussion sections
o Dicussion section: must include identified heterogeneity and important moderators, future
research directions, and theoretical relevance.
Checklist for qualitative meta-analysis reporting standards (Levitt et al., 2018)
Missing data
o Should provide sufficient detail and justification to inform readers about how missing
information was handled
missing reliability or range restriction values may be estimated using an artifact
distribution (Schmidt & Hunter, 2015)
missing elements of a correlation matrix may be imputed in meta-analytic structural
equation modeling (Cheung, 2000)
o ensure that meta-analysts handle missingness consistently and appropriately
o primary studies provided incomplete information (Higgins & Thomas, 2019) read later
Reference
Aguinis, H., Gottfredson, R. K., & Joo, H. (2013). Best-practice recommendations for defining, identifying,
and handling outliers [outlier]. Organizational Research Methods, 16(2), 270-301.
Andersson, U., Cuervo-Cazurra, A., & Nielsen, B. B. (2014). From the Editors: Explaining interaction effects
within and across levels of analysis. Journal of International Business Studies, 45(9), 1063-1071.
https://doi.org/10.1057/jibs.2014.50
DeSimone, J. A., Brannick, M. T., O’Boyle, E. H., & Ryu, J. W. (2021). Recommendations for reviewing meta-
analyses in organizational research. Organizational Research Methods, 24(4), 694-717.
Gusenbauer, M., & Haddaway, N. R. (2020). Which academic search systems are suitable for systematic
reviews or meta‐analyses? Evaluating retrieval qualities of Google Scholar, PubMed, and 26 other
resources. Research Synthesis Methods, 11(2), 181-217.
Hansen, C., Steinmetz, H., & Block, J. (2022). How to conduct a meta-analysis in eight steps: a practical
guide. Management Review Quarterly, 72(1), 1-19. https://doi.org/10.1007/s11301-021-00247-4
Harrison, J. S., Banks, G. C., Pollack, J. M., O’Boyle, E. H., & Short, J. (2017). Publication bias in strategic
management research. Journal of Management, 43(2), 400-425.
Hedges, L. V., Higgins, J. P. T., Rothstein, H. R., & Borenstein, M. (2009). Introduction to meta-analysis.
Chichester, UK.
Hunter, J. E., & Schmidt, F. L. (2004). Methods of meta-analysis: Correcting error and bias in research
findings (2nd ed.) [key]. Sage.
Kepes, S., Banks, G. C., Mcdaniel, M., & Whetzel, D. L. (2012). Publication Bias in the Organizational
Sciences. Organizational Research Methods, 15(4), 624-662.
https://doi.org/10.1177/1094428112452760
Levitt, H. M., Bamberg, M., Creswell, J. W., Frost, D. M., Josselson, R., & Suárez-Orozco, C. (2018). Journal
article reporting standards for qualitative primary, qualitative meta-analytic, and mixed methods
research in psychology: The APA Publications and Communications Board task force report [qua].
American Psychologist, 73(1), 26.
Lipsey, M. W., & Wilson, D. B. (2001). Practical meta-analysis. SAGE publications, Inc.
Marcus, J., & Le, H. (2013). Interactive effects of levels of individualism–collectivism on cooperation: A
meta‐analysis. Journal of Organizational Behavior, 34(6), 813-834. https://doi.org/10.1002/job.1875
Moher, D., Liberati, A., Tetzlaff, J., Altman, D. G., & Prisma Group*, t. (2009). Preferred reporting items for
systematic reviews and meta-analyses: the PRISMA statement [PRISMA]. Annals of internal medicine,
151(4), 264-269.
Murphy, K. R. R., Craig J. (2017). Mend It or End It: Redirecting the Search for Interactions in the
Organizational Sciences. Organizational Research Methods, 20(4), 549-573.
https://doi.org/10.1177/1094428115625322
Page, M. J., McKenzie, J. E., Bossuyt, P. M., Boutron, I., Hoffmann, T. C., Mulrow, C. D., Shamseer, L., Tetzlaff,
J. M., Akl, E. A., & Brennan, S. E. (2021). The PRISMA 2020 statement: an updated guideline for
reporting systematic reviews [PRISMA]. BMJ, 372.
Steel, P. D., & Kammeyer-Mueller, J. D. (2002). Comparing meta-analytic moderator estimation techniques
under realistic conditions. Journal of Applied Psychology, 87(1), 96.
Viechtbauer, W., & Cheung, M. W. L. (2010). Outlier and influence diagnostics for meta‐analysis [outlier].
Research Synthesis Methods, 1(2), 112-125.
Wood, J. A. (2008). Methodology for dealing with duplicate study effects in a meta-analysis. Organizational
Research Methods, 11(1), 79-95.