Abstract
For valid scientific discoveries, it is fundamental to evaluate whether research findings are replicable across different settings. While large-scale replication projects across broad research topics are not feasible, systematic reviews and meta-analyses (SRMAs) offer viable alternatives to assess repli-cability. Due to subjective inclusion and exclusion of studies, SRMAs may contain nonreplicable study findings. However, there is no consensus on rig-orous methods to assess the replicability of SRMAs or to explore sources of nonreplicability. Nonreplicability is often misconceived as high hetero-geneity. This article introduces a new measure, the externally standardized residuals from a leave-m-studies-out procedure, to quantify replicability. It not only measures the impact of nonreplicability from unknown sources on the conclusion of an SRMA but also differentiates nonreplicability from het-erogeneity. A new test statistic for replicability is derived. We explore its asymptotic properties and use extensive simulations and real data to illus-trate this measure’s performance. We conclude that replicability should be routinely assessed for all SRMAs and recommend sensitivity analyses, once nonreplicable study results are identified in an SRMA.
Original language | English (US) |
---|---|
Pages (from-to) | 664-682 |
Number of pages | 19 |
Journal | Annals of Applied Statistics |
Volume | 18 |
Issue number | 1 |
DOIs | |
State | Published - Mar 2024 |
Bibliographical note
Publisher Copyright:© Institute of Mathematical Statistics, 2024.
Keywords
- Externally standardized residual
- heterogeneity
- meta-analysis
- replicability
- statistical power
- systematic review