Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2023 Aug 15;4(8):101131.
doi: 10.1016/j.xcrm.2023.101131. Epub 2023 Jul 24.

RadioLOGIC, a healthcare model for processing electronic health records and decision-making in breast disease

Affiliations

RadioLOGIC, a healthcare model for processing electronic health records and decision-making in breast disease

Tianyu Zhang et al. Cell Rep Med. .

Abstract

Digital health data used in diagnostics, patient care, and oncology research continue to accumulate exponentially. Most medical information, and particularly radiology results, are stored in free-text format, and the potential of these data remains untapped. In this study, a radiological repomics-driven model incorporating medical token cognition (RadioLOGIC) is proposed to extract repomics (report omics) features from unstructured electronic health records and to assess human health and predict pathological outcome via transfer learning. The average accuracy and F1-weighted score for the extraction of repomics features using RadioLOGIC are 0.934 and 0.934, respectively, and 0.906 and 0.903 for the prediction of breast imaging-reporting and data system scores. The areas under the receiver operating characteristic curve for the prediction of pathological outcome without and with transfer learning are 0.912 and 0.945, respectively. RadioLOGIC outperforms cohort models in the capability to extract features and also reveals promise for checking clinical diagnoses directly from electronic health records.

Keywords: artificial intelligence; breast cancer; decision support; digital health data; electronic health records; radiology; repomics.

PubMed Disclaimer

Conflict of interest statement

Declaration of interests The authors declare no competing interests.

Figures

None
Graphical abstract
Figure 1
Figure 1
Characteristics of all radiological report content for the training cohort and the independent test cohort in this study (A) Training cohort. (B) Independent test cohort. A “0” means not applicable or missing. For margin: 1 for irregular, 2 for circumscribed. For shape: 1 for irregular, 2 for oval/round. Density category A (1): the breasts are almost entirely fatty. Density category B (2): there are scattered areas of fibroglandular density. Density category C (3): heterogeneously dense. Density category D (4): extremely dense. See also Table S1.
Figure 2
Figure 2
Visualizations of words and sentence (A) Word cloud based on all radiological reports. (B) Visualization of word co-occurrence. (C) Association of top 50 co-occurrence words. (D) Associations between words in a given report after pre-training. (E) The correlation between the selected word and other words in the report. See also Figures S1–S3 and Table S2.
Figure 3
Figure 3
Metrics details for extracting repomics features in the independent test cohort (A) Accuracy. (B) Micro F1 score. (C) Macro F1 score. (D) Weighted F1 score. The black vertical lines in the graph represent 95% confidence intervals.
Figure 4
Figure 4
Examples of repomics feature extraction from corresponding images and radiological reports (A) Mammography. (B) Ultrasound. (C) MRI. “∗∗∗” in the radiological report indicates the patient’s private information. See also Figures S4–S7.
Figure 5
Figure 5
Prediction results for downstream tasks (A) Confusion matrix results for predicting BI-RADS scores in the independent test cohort using RadioLOGIC without transfer learning. (B) Confusion matrix results for predicting BI-RADS scores in the independent test cohort using RadioLOGIC via transfer learning. (C) Receiver operating characteristic curves for predicting pathological outcome in the independent test cohort using RNN. (D) Receiver operating characteristic curves for predicting pathological outcome in the independent test cohort using RadioLOGIC. The 95% confidence intervals are shown as a shaded area for the ROC curve. ATT, attention mechanism; BI-RADS, breast imaging-reporting and data system; RadioLOGIC, radiological repomics-driven model incorporating medical token cognition; RNN, recurrent neural networks; TF, transfer learning. See also Table S3.

References

    1. Sung H., Ferlay J., Siegel R.L., Laversanne M., Soerjomataram I., Jemal A., Bray F. Global cancer statistics 2020: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA. Cancer J. Clin. 2021;71:209–249. - PubMed
    1. Sechopoulos I., Teuwen J., Mann R. Artificial intelligence for breast cancer detection in mammography and digital breast tomosynthesis: State of the art. Semin. Cancer Biol. 2021;72:214–225. doi: 10.1016/j.semcancer.2020.06.002. - DOI - PubMed
    1. Xue P., Wang J., Qin D., Yan H., Qu Y., Seery S., Jiang Y., Qiao Y. Deep learning in image-based breast and cervical cancer detection: a systematic review and meta-analysis. NPJ Digit. Med. 2022;5:19. - PMC - PubMed
    1. He J., Baxter S.L., Xu J., Xu J., Zhou X., Zhang K. The practical implementation of artificial intelligence technologies in medicine. Nat. Med. 2019;25:30–36. - PMC - PubMed
    1. Abul-Husn N.S., Kenny E.E. Personalized medicine and the power of electronic health records. Cell. 2019;177:58–69. - PMC - PubMed

Publication types