An evaluation of the performance of stopping rules in AI-aided screening for psychological meta-analytical research
- PMID: 39412090
- DOI: 10.1002/jrsm.1762
An evaluation of the performance of stopping rules in AI-aided screening for psychological meta-analytical research
Abstract
Several AI-aided screening tools have emerged to tackle the ever-expanding body of literature. These tools employ active learning, where algorithms sort abstracts based on human feedback. However, researchers using these tools face a crucial dilemma: When should they stop screening without knowing the proportion of relevant studies? Although numerous stopping rules have been proposed to guide users in this decision, they have yet to undergo comprehensive evaluation. In this study, we evaluated the performance of three stopping rules: the knee method, a data-driven heuristic, and a prevalence estimation technique. We measured performance via sensitivity, specificity, and screening cost and explored the influence of the prevalence of relevant studies and the choice of the learning algorithm. We curated a dataset of abstract collections from meta-analyses across five psychological research domains. Our findings revealed performance differences between stopping rules regarding all performance measures and variations in the performance of stopping rules across different prevalence ratios. Moreover, despite the relatively minor impact of the learning algorithm, we found that specific combinations of stopping rules and learning algorithms were most effective for certain prevalence ratios of relevant abstracts. Based on these results, we derived practical recommendations for users of AI-aided screening tools. Furthermore, we discuss possible implications and offer suggestions for future research.
Keywords: literature screening; machine learning; meta‐analysis; stopping rules; systematic reviews.
© 2024 The Author(s). Research Synthesis Methods published by John Wiley & Sons Ltd.
References
REFERENCES
-
- Elliott JH, Synnot A, Turner T, et al. Living systematic review: 1. Introduction—the why, what, when, and how. J Clin Epidemiol. 2017;91:23‐30. doi:10.1016/j.jclinepi.2017.08.010
-
- Borah R, Brown AW, Capers PL, Kaiser KA. Analysis of the time and workers needed to conduct systematic reviews of medical interventions using data from the PROSPERO registry. BMJ Open. 2017;7(2):e012545. doi:10.1136/bmjopen‐2016‐012545
-
- Marshall IJ, Wallace BC. Toward systematic review automation: a practical guide to using machine learning tools in research synthesis. Syst Rev. 2019;8(1):163. doi:10.1186/s13643‐019‐1074‐9
-
- Pham B, Bagheri E, Rios P, et al. Improving the conduct of systematic reviews: a process mining perspective. J Clin Epidemiol. 2018;103:101‐111. doi:10.1016/j.jclinepi.2018.06.011
-
- Cooper HM, Hedges LV, Valentine JC, eds. The Handbook of Research Synthesis and Meta‐Analysis. 2nd ed. Russell Sage Foundation; 2009.
MeSH terms
Grants and funding
LinkOut - more resources
Full Text Sources
