Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
Multicenter Study
. 2025 Jun 16;25(1):75.
doi: 10.1186/s40644-025-00892-y.

Whole-lesion-aware network based on freehand ultrasound video for breast cancer assessment: a prospective multicenter study

Affiliations
Multicenter Study

Whole-lesion-aware network based on freehand ultrasound video for breast cancer assessment: a prospective multicenter study

Jie Han et al. Cancer Imaging. .

Abstract

Background: The clinical application of artificial intelligence (AI) models based on breast ultrasound static images has been hindered in real-world workflows due to operator-dependence of standardized image acquisition and incomplete view of breast lesions on static images. To better exploit the real-time advantages of ultrasound and more conducive to clinical application, we proposed a whole-lesion-aware network based on freehand ultrasound video (WAUVE) scanning in an arbitrary direction for predicting overall breast cancer risk score.

Methods: The WAUVE was developed using 2912 videos (2912 lesions) of 2771 patients retrospectively collected from May 2020 to August 2022 in two hospitals. We compared the diagnostic performance of WAUVE with static 2D-ResNet50 and dynamic TimeSformer models in the internal validation set. Subsequently, a dataset comprising 190 videos (190 lesions) from 175 patients prospectively collected from December 2022 to April 2023 in two other hospitals, was used as an independent external validation set. A reader study was conducted by four experienced radiologists on the external validation set. We compared the diagnostic performance of WAUVE with the four experienced radiologists and evaluated the auxiliary value of model for radiologists.

Results: The WAUVE demonstrated superior performance compared to the 2D-ResNet50 model, while similar to the TimeSformer model. In the external validation set, WAUVE achieved an area under the receiver operating characteristic curve (AUC) of 0.8998 (95% CI = 0.8529-0.9439), and showed a comparable diagnostic performance to that of four experienced radiologists in terms of sensitivity (97.39% vs. 98.48%, p = 0.36), specificity (49.33% vs. 50.00%, p = 0.92), and accuracy (78.42% vs.79.34%, p = 0.60). With the WAUVE model assistance, the average specificity of four experienced radiologists was improved by 6.67%, and higher consistency was achieved (from 0.807 to 0.838).

Conclusion: The WAUVE based on non-standardized ultrasound scanning demonstrated excellent performance in breast cancer assessment which yielded outcomes similar to those of experienced radiologists, indicating the clinical application of the WAUVE model promising.

Keywords: Artificial intelligence; Breast neoplasms; Deep learning; Diagnosis; Ultrasonography; Video.

PubMed Disclaimer

Conflict of interest statement

Declarations. Ethics approval and consent to participate: This multicenter study was divided into two parts: model development using a retrospective dataset and model validation using a prospective dataset. The retrospective study was approved by Ethics Committee of National Cancer Center/Cancer Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College (No. 22/386-3588) with a waiver granted for the requirement of informed consent. The prospective study was approved by Ethics Committee of National Cancer Center/Cancer Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College (No. 21/328-2999), and informed consent was obtained from each patient or their guardian before the US examination. Consent for publication: Not applicable. Competing interests: The authors declare no competing interests.

Figures

Fig. 1
Fig. 1
Overview of the retrospective development and prospective external validation dataset workflows. The development dataset was collected in two hospitals before August 2022, while the external validation dataset was collected in two other hospitals after development of the deep learning model. Abbreviations: NCC, National Cancer Center; PUMCH, Peking Union Medical College Hospital
Fig. 2
Fig. 2
The schematic architecture of the whole-lesion-aware network for freehand ultrasound video(WAUVE). For a given ultrasound video, (a) the Faster R-CNN with ResNet-34 backbone firstly extracts regions-of-interest (ROIs) from each frame. Then ROIs are cropped to generate video clips for each video, where each adjacent eight frames are grouped as a video clip. Subsequently, (b) Inflated 3D ConvNet (I3D) with ResNet-50 backbone performs malignancy probability prediction for each video clip. Finally, the overall cancer risk score is outputted in the form of the average of the results of all video clips
Fig. 3
Fig. 3
Receiver operating characteristic curves for different DL models in the internal validation dataset. Abbreviations: AUC, area under the receiver operating characteristic curve; DL, deep learning
Fig. 4
Fig. 4
Diagnostic performance comparison between WAUVE model and four radiologists. Receiver operating characteristic (ROC) curves and points are used to depict the performance of the model and four radiologists, respectively. Additionally, the mean performance of four radiologists is presented by error bars with 95% confidence intervals (CIs), which were calculated based on 9999 bootstraps of the data. Green and purple points represent the performance of radiologists without and with AI assistance, respectively
Fig. 5
Fig. 5
Consistency between different radiologists before (a) and after (b) AI assistance. Based on the color bar on the right, the color block darkens with increasing consistency, and the number in the color block area represents the Kappa value between two corresponding radiologists. With AI assistance, the consistency between radiologists was higher than that before AI assistance, and the Kappa values between different radiologists were all improved to different degrees
Fig. 6
Fig. 6
Examples of Qualitative illustration and heatmap illustration of our WAUVE model. We present representative frames from the input video and the corresponding heatmaps. The green bounding boxes represent the detected ROI in each frame. For each ROI, the cancer risk score of the video clip (Clip N: prob = X) is presented above the bounding box. In the heatmap, the red regions contribute to high malignancy probability and the blue regions to benign. a The WAUVE model obtained an overall cancer risk score of 0.643, with mostly red-colored regions suggesting a malignant lesion. All four radiologists in the reader study classified this lesion as BI-RADS 4 C, and the final histopathological diagnosis was a malignancy. b The WAUVE model obtained an overall cancer risk score of 0.111, with mostly blue regions suggesting a benign lesion. One radiologist assigned this lesion as BI-RADS 4a and another as 4b, indicating mild to moderate suspicion for malignancy. However, the final diagnosis was a benign fibroadenoma

Similar articles

References

    1. Bray F, Laversanne M, Sung H, Ferlay J, Siegel RL, Soerjomataram I, et al. Global cancer statistics 2022: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA Cancer J Clin. 2024;74:229–63. - PubMed
    1. Kexin S, Lin L, Rongshou Z, Siwei Z, Hongmei Z, Shaoming W, et al. Trends in incidence rates, mortality rates, and age-period-cohort effects of female breast cancer — China, 2003–2017. China CDC Weekly. 2023;5:340–6. - PMC - PubMed
    1. McBee MP, Awan OA, Colucci AT, Ghobadi CW, Kadom N, Kansagra AP, et al. Deep Learning in radiology. Acad Radiol. 2018;25:1472–80. - PubMed
    1. Avanzo M, Wei L, Stancanello J, Vallières M, Rao A, Morin O, et al. Machine and deep learning methods for radiomics. Med Phys. 2020;47:e185–202. - PMC - PubMed
    1. Chan HP, Samala RK, Hadjiiski LM, Zhou C. Deep learning in medical image analysis. Adv Exp Med Biol. 2020;1213:3–21. - PMC - PubMed

Publication types

LinkOut - more resources