Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2022 Feb:5:189-200.
doi: 10.5220/0010903300003123.

The h-ANN Model: Comprehensive Colonoscopy Concept Compilation Using Combined Contextual Embeddings

Affiliations

The h-ANN Model: Comprehensive Colonoscopy Concept Compilation Using Combined Contextual Embeddings

Shorabuddin Syed et al. Biomed Eng Syst Technol Int Jt Conf BIOSTEC Revis Sel Pap. 2022 Feb.

Abstract

Colonoscopy is a screening and diagnostic procedure for detection of colorectal carcinomas with specific quality metrics that monitor and improve adenoma detection rates. These quality metrics are stored in disparate documents i.e., colonoscopy, pathology, and radiology reports. The lack of integrated standardized documentation is impeding colorectal cancer research. Clinical concept extraction using Natural Language Processing (NLP) and Machine Learning (ML) techniques is an alternative to manual data abstraction. Contextual word embedding models such as BERT (Bidirectional Encoder Representations from Transformers) and FLAIR have enhanced performance of NLP tasks. Combining multiple clinically-trained embeddings can improve word representations and boost the performance of the clinical NLP systems. The objective of this study is to extract comprehensive clinical concepts from the consolidated colonoscopy documents using concatenated clinical embeddings. We built high-quality annotated corpora for three report types. BERT and FLAIR embeddings were trained on unlabeled colonoscopy related documents. We built a hybrid Artificial Neural Network (h-ANN) to concatenate and fine-tune BERT and FLAIR embeddings. To extract concepts of interest from three report types, 3 models were initialized from the h-ANN and fine-tuned using the annotated corpora. The models achieved best F1-scores of 91.76%, 92.25%, and 88.55% for colonoscopy, pathology, and radiology reports respectively.

Keywords: Clinical Concept Extraction; Colonoscopy; Deep Learning; Natural Language Processing; Word Embeddings.

PubMed Disclaimer

Figures

Figure 1:
Figure 1:
Colonoscopy taxonomy depicting clinical entities and their classifications. Colonoscopy reports were annotated for entities mentioned in the taxonomy.
Figure 2:
Figure 2:
Pathology taxonomy depicting clinical entities and their classifications. Pathology reports were annotated for entities mentioned in the taxonomy.
Figure 3:
Figure 3:
Radiology imaging taxonomy depicting clinical entities and their classifications. Radiology reports were annotated for entities mentioned in the taxonomy.
Figure 4:
Figure 4:
Workflow depicting training of language models, concatenating embeddings, instantiating and fine-tuning h-ANN models to extract clinical concepts from colonoscopy related documents. GI: Gastroenterology, h-ANN: Hybrid artificial neural network.
Figure 5:
Figure 5:
The h-ANN architecture depicting embedding, Bi-LSTM, and CRF layers. Concatenated BERT and FLAIR embeddings are given as input features to the Bi-LSTM layer. BERT: Bidirectional Encoder Representations from Transformers.
Figure 6:
Figure 6:
Training curves for h-ANNpath, h-ANNcol, and h-ANNrad models. F1 score on the test set was measured after completion of each epochs.

References

    1. Abadi M, Agarwal A, Barham P, Brevdo E, Chen Z, Citro C, … Zheng XJA. (2016). TensorFlow: Large-Scale Machine Learning on Heterogeneous Distributed Systems. abs/1603.04467.
    1. Abdalla M, Abdalla M, Rudzicz F, & Hirst G (2020). Using word embeddings to improve the privacy of clinical notes. Journal of the American Medical Informatics Association, 27(6), 901–907. doi:10.1093/jamia/ocaa038 - DOI - PMC - PubMed
    1. Akbik A, Bergmann T, Blythe D, Rasul K, Schweter S, & Vollgraf R (2019). FLAIR: An Easy-to-Use Framework for State-of-the-Art NLP. 10.18653/v1/n19-4010 - DOI
    1. Akbik A, Blythe D, & Vollgraf R (2018). Contextual String Embeddings for Sequence Labeling. Santa Fe, New Mexico, USA: Association for Computational Linguistics.
    1. Alsentzer E, Murphy J, Boag W, Weng W-H, Jindi D, Naumann T, & McDermott M (2019). Publicly Available Clinical BERT Embeddings. Minneapolis, Minnesota, USA: Association for Computational Linguistics.

LinkOut - more resources