Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2023 Dec:164:1-8.
doi: 10.1016/j.jclinepi.2023.10.007. Epub 2023 Oct 21.

Crowd-sourcing and automation facilitated the identification and classification of randomized controlled trials in a living review

Affiliations
Free article

Crowd-sourcing and automation facilitated the identification and classification of randomized controlled trials in a living review

Mohammed Mujaab Kamso et al. J Clin Epidemiol. 2023 Dec.
Free article

Abstract

Objectives: To evaluate an approach using automation and crowdsourcing to identify and classify randomized controlled trials (RCTs) for rheumatoid arthritis (RA) in a living systematic review (LSR).

Methods: Records from a database search for RCTs in RA were screened first by machine learning and Cochrane Crowd to exclude non-RCTs, then by trainee reviewers using a Population, Intervention, Comparison, and Outcome (PICO) annotator platform to assess eligibility and classify the trial to the appropriate review. Disagreements were resolved by experts using a custom online tool. We evaluated the efficiency gains, sensitivity, accuracy, and interrater agreement (kappa scores) between reviewers.

Results: From 42,452 records, machine learning and Cochrane Crowd excluded 28,777 (68%), trainee reviewers excluded 4,529 (11%), and experts excluded 7,200 (17%). The 1,946 records eligible for our LSR represented 220 RCTs and included 148/149 (99.3%) of known eligible trials from prior reviews. Although excluded from our LSRs, 6,420 records were classified as other RCTs in RA to inform future reviews. False negative rates among trainees were highest for the RCT domain (12%), although only 1.1% of these were for the primary record. Kappa scores for two reviewers ranged from moderate to substantial agreement (0.40-0.69).

Conclusion: A screening approach combining machine learning, crowdsourcing, and trainee participation substantially reduced the screening burden for expert reviewers and was highly sensitive.

Keywords: Automation; Crowdsourcing; Living systematic reviews; Machine learning; Randomized controlled trials (RCTs); Rheumatoid arthritis; Systematic reviews.

PubMed Disclaimer

Conflict of interest statement

Declaration of competing interest None.

Publication types

LinkOut - more resources