Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2021 Mar 1;28(3):504-515.
doi: 10.1093/jamia/ocaa261.

Can reproducibility be improved in clinical natural language processing? A study of 7 clinical NLP suites

Affiliations

Can reproducibility be improved in clinical natural language processing? A study of 7 clinical NLP suites

William Digan et al. J Am Med Inform Assoc. .

Abstract

Background: The increasing complexity of data streams and computational processes in modern clinical health information systems makes reproducibility challenging. Clinical natural language processing (NLP) pipelines are routinely leveraged for the secondary use of data. Workflow management systems (WMS) have been widely used in bioinformatics to handle the reproducibility bottleneck.

Objective: To evaluate if WMS and other bioinformatics practices could impact the reproducibility of clinical NLP frameworks.

Materials and methods: Based on the literature across multiple researcho fields (NLP, bioinformatics and clinical informatics) we selected articles which (1) review reproducibility practices and (2) highlight a set of rules or guidelines to ensure tool or pipeline reproducibility. We aggregate insight from the literature to define reproducibility recommendations. Finally, we assess the compliance of 7 NLP frameworks to the recommendations.

Results: We identified 40 reproducibility features from 8 selected articles. Frameworks based on WMS match more than 50% of features (26 features for LAPPS Grid, 22 features for OpenMinted) compared to 18 features for current clinical NLP framework (cTakes, CLAMP) and 17 features for GATE, ScispaCy, and Textflows.

Discussion: 34 recommendations are endorsed by at least 2 articles from our selection. Overall, 15 features were adopted by every NLP Framework. Nevertheless, frameworks based on WMS had a better compliance with the features.

Conclusion: NLP frameworks could benefit from lessons learned from the bioinformatics field (eg, public repositories of curated tools and workflows or use of containers for shareability) to enhance the reproducibility in a clinical setting.

Keywords: containerization; meaningful use; natural language processing; reproducibility of results; workflow.

PubMed Disclaimer

Figures

Figure 1.
Figure 1.
Reproducibility articles sorted by level of analysis and research fields. The scope category as either tool or WMS is also shown.
Figure 2.
Figure 2.
Classification of recommendation at tool and pipeline level. 21 recommendations are applicable both to tools and workflows, 12 to workflows only, and 7 to tools only.

References

    1. Baker M. 1,500 scientists lift the lid on reproducibility. Nature News 2016; 533 (7604): 452–4. - PubMed
    1. Cohen-Boulakia S, Belhajjame K, Collin O, et al.Scientific workflows for computational reproducibility in the life sciences: status, challenges and opportunities. Future Gen Comput Syst 2017; 75: 284–98.
    1. Wilkinson MD, Dumontier M, Aalbersberg IJ, et al.The FAIR Guiding Principles for scientific data management and stewardship. Sci Data 2016; 3 (1): 1–9. - PMC - PubMed
    1. Collberg C, Proebsting TA.. Repeatability in computer systems research. Commun ACM 2016; 59 (3): 62–9.
    1. Benureau FCY, Rougier NP.. Re-run, repeat, reproduce, reuse, replicate: transforming code into scientific contributions. Front Neuroinform 2018; 11: 1–8. - PMC - PubMed

Publication types

MeSH terms