Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2022 Dec 2;13(1):7456.
doi: 10.1038/s41467-022-35007-9.

Deciphering clinical abbreviations with a privacy protecting machine learning system

Affiliations

Deciphering clinical abbreviations with a privacy protecting machine learning system

Alvin Rajkomar et al. Nat Commun. .

Abstract

Physicians write clinical notes with abbreviations and shorthand that are difficult to decipher. Abbreviations can be clinical jargon (writing "HIT" for "heparin induced thrombocytopenia"), ambiguous terms that require expertise to disambiguate (using "MS" for "multiple sclerosis" or "mental status"), or domain-specific vernacular ("cb" for "complicated by"). Here we train machine learning models on public web data to decode such text by replacing abbreviations with their meanings. We report a single translation model that simultaneously detects and expands thousands of abbreviations in real clinical notes with accuracies ranging from 92.1%-97.1% on multiple external test datasets. The model equals or exceeds the performance of board-certified physicians (97.6% vs 88.7% total accuracy). Our results demonstrate a general method to contextually decipher abbreviations and shorthand that is built without any privacy-compromising data.

PubMed Disclaimer

Conflict of interest statement

All authors are employed by Google as indicated by the affiliation. Google has filed a provisional patent application 63/269,420 that is related to this article.

Figures

Fig. 1
Fig. 1. Overview of task formulation and comparison against traditional approaches.
A Task formulation: The input to our model is a string that may or may not contain medical abbreviations. We trained a model to output a corresponding string in which all abbreviations are simultaneously detected and expanded. If the input string does not contain an abbreviation, the model will output the original string. B Traditional vs. our approach: The traditional approach for medical abbreviation disambiguation divides the task into separate steps. First, a separate component identifies which tokens of an input string are abbreviations. Second, each abbreviation’s surrounding context is processed by an abbreviation-specific model that outputs scores associated with a closed set of possible expansions of the abbreviation. This not only requires training and deploying thousands of models, but each model also cannot benefit from the inductive bias of the other models, especially as it relates to any other abbreviations that may appear in the surrounding context of a given abbreviation. Finally, each expansion must be inserted back into the original snippet, which may require additional post-processing to maintain grammatical correctness. In our approach, a single end-to-end process performs all of these steps in parallel: the detection of abbreviations, the expansion of all abbreviations such that their likelihood is maximized in a mutually consistent manner, and the generation of a grammatically consistent “decoded” piece of text.
Fig. 2
Fig. 2. Overview of web-scale reverse substitution.
Web scale reverse substitution: Large language models are often pre-trained on public web data to carry out general self-supervised tasks. To further train (i.e. fine-tune) a model to decipher clinical abbreviations, we use the following procedure to generate an additional dataset as shown conceptually in this panel. We take public web pages and identify words or phrases that have corresponding abbreviations (the colored boxes on the left-hand side) in the dictionary released along with this manuscript. We substitute the abbreviation (the colored dots) to generate input text to train our model for the decoding task. If a word/phrase has more than one abbreviation (e.g. “atrial fibrillation” could be “af’ or “afib”), we randomly pick one. Given the size of the web-corpus and the imbalance of expansions (e.g. “patient” is found orders of magnitude more often than “posterior tibialis”), a simple “find and replace” is problematic because it creates a large, imbalanced dataset. Instead, our algorithm, which we call web-scale reverse substitution, downsamples frequent expansions to derive a more balanced dataset from a web-sized corpus of thousands of expansions.
Fig. 3
Fig. 3. Effect of model size and inference type on performance.
Model size and inference type influence key model metrics on the synthetic test set. Each point reflects a T5 model with identical pre-training, MC-WSRS fine-tuning, and evaluation on a synthetic dataset of medical snippets. Detection recall decreases as the model size increases. However, performance is substantially and statistically improved with inference chaining techniques such as iterative inference and elicitive inference. The inference types and model sizes do not significantly affect detection precision (percentage of the text identified as abbreviations that are actually abbreviations), and model size improves expansion accuracy (percentage of expansions with clinically equivalent meanings). Total accuracy, which we define as detection recall multiplied by expansion accuracy, is highest for the model with the most parameters with elicitive inference. n = 400 bootstrap samples of the 302 synthetic snippets, each of which contains a different collection of abbreviations. Point estimates from the original sample and 95% confidence intervals were calculated using reporting the 2.5 and 97.5 percentile values for each metric across the samples.

References

    1. Leveille SG, et al. Patients evaluate visit notes written by their clinicians: a mixed methods investigation. J. Gen. Intern. Med. 2020;35:3510–3516. doi: 10.1007/s11606-020-06014-7. - DOI - PMC - PubMed
    1. Federal Rules Mandating Open Notes. 2022. https://www.opennotes.org/onc-federal-rule/.
    1. Grossman Liu L, et al. Effect of expansion of abbreviations and acronyms on patient comprehension of their health records: a randomized clinical trial. JAMA Netw. Open. 2022;5:e2212320. doi: 10.1001/jamanetworkopen.2022.12320. - DOI - PMC - PubMed
    1. Chemali, M., Hibbert, E. J. & Sheen, A General practitioner understanding of abbreviations used in hospital discharge letters. Med. J. Aust.203, 147, 147e.1–4. (2015). - PubMed
    1. Chemali M, Hibbert EJ, Sheen A. General practitioner understanding of abbreviations used in hospital discharge letters. Med. J. Aust. 2015;203:147. doi: 10.5694/mja15.00224. - DOI - PubMed