Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2022 Aug 4;17(8):e0270595.
doi: 10.1371/journal.pone.0270595. eCollection 2022.

Multi-label classification of symptom terms from free-text bilingual adverse drug reaction reports using natural language processing

Affiliations

Multi-label classification of symptom terms from free-text bilingual adverse drug reaction reports using natural language processing

Sitthichok Chaichulee et al. PLoS One. .

Abstract

Allergic reactions to medication range from mild to severe or even life-threatening. Proper documentation of patient allergy information is critical for safe prescription, avoiding drug interactions, and reducing healthcare costs. Allergy information is regularly obtained during the medical interview, but is often poorly documented in electronic health records (EHRs). While many EHRs allow for structured adverse drug reaction (ADR) reporting, a free-text entry is still common. The resulting information is neither interoperable nor easily reusable for other applications, such as clinical decision support systems and prescription alerts. Current approaches require pharmacists to review and code ADRs documented by healthcare professionals. Recently, the effectiveness of machine algorithms in natural language processing (NLP) has been widely demonstrated. Our study aims to develop and evaluate different NLP algorithms that can encode unstructured ADRs stored in EHRs into institutional symptom terms. Our dataset consists of 79,712 pharmacist-reviewed drug allergy records. We evaluated three NLP techniques: Naive Bayes-Support Vector Machine (NB-SVM), Universal Language Model Fine-tuning (ULMFiT), and Bidirectional Encoder Representations from Transformers (BERT). We tested different general-domain pre-trained BERT models, including mBERT, XLM-RoBERTa, and WanchanBERTa, as well as our domain-specific AllergyRoBERTa, which was pre-trained from scratch on our corpus. Overall, BERT models had the highest performance. NB-SVM outperformed ULMFiT and BERT for several symptom terms that are not frequently coded. The ensemble model achieved an exact match ratio of 95.33%, a F1 score of 98.88%, and a mean average precision of 97.07% for the 36 most frequently coded symptom terms. The model was then further developed into a symptom term suggestion system and achieved a Krippendorff's alpha agreement coefficient of 0.7081 in prospective testing with pharmacists. Some degree of automation could both accelerate the availability of allergy information and reduce the efforts for human coding.

PubMed Disclaimer

Conflict of interest statement

The authors have declared that no competing interests exist.

Figures

Fig 1
Fig 1. Flowchart outlines the steps involved in our study.
Drug allergy records in the Songklanagarind Hospital’s EHR from October 2001 to July 2020 were extracted for developing algorithms. Only the records that had been reviewed by pharmacists were considered. We then trained NB-SVM, ULMFiT, and BERT-based models to map the unstructured allergy description to the institutional symptom terms. The ensemble model was then used for evaluation with pharmacists via a web application in a simulated EHR environment.
Fig 2
Fig 2. GUI for document a new drug allergy record.
GUI for documenting a drug allergy record in the Songklanagarind Hospital’s EHR. The user interface was originally in Thai and was translated to English for ease of the reader.
Fig 3
Fig 3. Diagrams outline the steps involved in our methods.
(A) Data preparation included word segmentation and tokenization, with each algorithm using different method. (B) NB-SVM involved training multiple pipelines for Naive Bayes feature extraction and SVM classification. (C) ULMFiT involved fine-tuning the pre-trained LM with our target-domain allergy corpus and fine-tuning a classifier for our multi-label classification task. (D) BERT involves fine-tuning a classifier for our multi-label classification task. This study evaluated three pre-trained general-domain BERT models and one target-domain BERT model pre-trained on our allergy corpus.

Similar articles

Cited by

References

    1. Khan DA, Solensky R. Drug Allergy. Journal of Allergy and Clinical Immunology. 2010;125(2):S126–S137. doi: 10.1016/j.jaci.2009.10.028 - DOI - PubMed
    1. Thong BYH, Tan TC. Epidemiology and Risk Factors for Drug Allergy. British Journal of Clinical Pharmacology. 2011. May;71(5):684–700. doi: 10.1111/j.1365-2125.2010.03774.x - DOI - PMC - PubMed
    1. Warrington R, Silviu-Dan F. Drug Allergy. Allergy, Asthma & Clinical Immunology. 2011;7(S1):S10. doi: 10.1186/1710-1492-7-S1-S10 - DOI - PMC - PubMed
    1. Greenberger PA. Drug Allergy. Allergy and Asthma Proceedings. 2019. Nov;40(6):474–479. doi: 10.2500/aap.2019.40.4275 - DOI - PubMed
    1. Epstein RH, St Jacques P, Stockin M, Rothman B, Ehrenfeld JM, Denny JC. Automated Identification of Drug and Food Allergies Entered Using Non-standard Terminology. Journal of the American Medical Informatics Association. 2013;20(5):962–968. doi: 10.1136/amiajnl-2013-001756 - DOI - PMC - PubMed