2017
DOI: 10.31219/osf.io/w89s5
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

New statistical metrics for multisite replication projects

Abstract: Increasing interest in replicability in the social sciences has engendered novel designs for replication projects in which multiple sites replicate an original study. At least 134 such "many-to-one" replications have been completed since 2014 or are currently ongoing. These designs have unique potential to help estimate whether the original study is statistically consistent with the replications and to re-assess the strength of evidence for the scientific effect of interest. However, existing statistical analy… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
15
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
6
1

Relationship

1
6

Authors

Journals

citations
Cited by 8 publications
(15 citation statements)
references
References 45 publications
0
15
0
Order By: Relevance
“…Under this situation, one could informally say that an effect is "replicable." When several replications are available, a more quantitative meta-analytic approach can be taken: an effect can be considered "replicable" when the meta-analytic ES estimate excludes zero and is consistent with the original ES point estimate (also replication scenario #1, see Panel A Figure 1; see also Mathur & VanderWeele, 2018).…”
Section: Statistical Interpretation: Meta-analytic Levelmentioning
confidence: 99%
“…Under this situation, one could informally say that an effect is "replicable." When several replications are available, a more quantitative meta-analytic approach can be taken: an effect can be considered "replicable" when the meta-analytic ES estimate excludes zero and is consistent with the original ES point estimate (also replication scenario #1, see Panel A Figure 1; see also Mathur & VanderWeele, 2018).…”
Section: Statistical Interpretation: Meta-analytic Levelmentioning
confidence: 99%
“…First, we will conduct a form of equivalence testing by assessing, for each study, the proportion of effect sizes in the potentially heterogeneous distribution underlying the replications that are greater than an effect size of 0 and greater than a small effect size (e.g., r = .10; Mathur & VanderWeele, 2017). We will also examine the observed replicability rate (defined as the proportion of replications for each study that estimated a statistically significant point estimate in the same direction as that of the original study) compared to the rate we would expect in ideal circumstances (e.g., those without selective analyses or reporting) following the guidelines of Mathur and VanderWeele (2017). Finally, we will examine the consistency of each set of replications (either separated by protocol or pooled) with the results of the original studies and RP:P replications, again following the guidelines of Mathur and VanderWeele (2017).…”
Section: Exploratory Analyses -Other Evaluations Of Replicabilitymentioning
confidence: 99%
“…We will also examine the observed replicability rate (defined as the proportion of replications for each study that estimated a statistically significant point estimate in the same direction as that of the original study) compared to the rate we would expect in ideal circumstances (e.g., those without selective analyses or reporting) following the guidelines of Mathur and VanderWeele (2017). Finally, we will examine the consistency of each set of replications (either separated by protocol or pooled) with the results of the original studies and RP:P replications, again following the guidelines of Mathur and VanderWeele (2017). ]…”
Section: Exploratory Analyses -Other Evaluations Of Replicabilitymentioning
confidence: 99%
“…Other considerations are also relevant to replication success. One can assess the extent to which the original study is statistically consistent with the replication study, in that its data are compatible with the same underlying population parameter that is unbiasedly estimated in the replications (Mathur and VanderWeele, ). This method assesses whether the replication and original studies’ point estimates are more distant than expected by chance, accounting for statistical uncertainty in both, and hence are conceptually different from Held's reverse Bayesian sensitivity analysis on the original study.…”
Section: Proportion Of Successful Replication Based On Different Defimentioning
confidence: 99%
“…This method assesses whether the replication and original studies’ point estimates are more distant than expected by chance, accounting for statistical uncertainty in both, and hence are conceptually different from Held's reverse Bayesian sensitivity analysis on the original study. For example, the consistency metric accommodates ‘non‐significant’ original studies and those with a priori plausible or implausible findings; it also applies readily to multisite replication studies, accounting for effect heterogeneity (Mathur and VanderWeele, ). With multisite replications, one can also reassess the strength of evidence for the effect under investigation by meta‐analytic methods to estimate the proportion of true effect sizes in the replications that are stronger than a chosen threshold of scientific importance (Mathur and VanderWeele, ).…”
Section: Proportion Of Successful Replication Based On Different Defimentioning
confidence: 99%