Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2023 Oct 13:28:100500.
doi: 10.1016/j.phro.2023.100500. eCollection 2023 Oct.

A network score-based metric to optimize the quality assurance of automatic radiotherapy target segmentations

Affiliations

A network score-based metric to optimize the quality assurance of automatic radiotherapy target segmentations

Roque Rodríguez Outeiral et al. Phys Imaging Radiat Oncol. .

Abstract

Background and purpose: Existing methods for quality assurance of the radiotherapy auto-segmentations focus on the correlation between the average model entropy and the Dice Similarity Coefficient (DSC) only. We identified a metric directly derived from the output of the network and correlated it with clinically relevant metrics for contour accuracy.

Materials and methods: Magnetic Resonance Imaging auto-segmentations were available for the gross tumor volume for cervical cancer brachytherapy (106 segmentations) and for the clinical target volume for rectal cancer external-beam radiotherapy (77 segmentations). The nnU-Net's output before binarization was taken as a score map. We defined a metric as the mean of the voxels in the score map above a threshold (λ). Comparisons were made with the mean and standard deviation over the score map and with the mean over the entropy map. The DSC, the 95th Hausdorff distance, the mean surface distance (MSD) and the surface DSC were computed for segmentation quality. Correlations between the studied metrics and model quality were assessed with the Pearson correlation coefficient (r). The area under the curve (AUC) was determined for detecting segmentations that require reviewing.

Results: For both tasks, our metric (λ = 0.30) correlated more strongly with the segmentation quality than the mean over the entropy map (for surface DSC, r > 0.65 vs. r < 0.60). The AUC was above 0.84 for detecting MSD values above 2 mm.

Conclusions: Our metric correlated strongly with clinically relevant segmentation metrics and detected segmentations that required reviewing, indicating its potential for automatic quality assurance of radiotherapy target auto-segmentations.

Keywords: Automatic segmentation; Cervical cancer; Confidence estimation; Quality assurance; Rectal cancer.

PubMed Disclaimer

Conflict of interest statement

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Figures

Fig. 1
Fig. 1
Workflow of the study design.
Fig. 2
Fig. 2
Difference in Pearson correlation coefficient (Δr) with the segmentation metrics between the HiS metric and the mean over the score map as a function of the parameter λ. The bold line is the average Δr among the five folds. The dashed lines represent the Δr for each of the five folds.
Fig. 3
Fig. 3
Scatter plots between the segmentation metrics and the HiS metric for the cervical cancer cohort (top) and the rectal cancer cohort (bottom). The translucent band corresponds to the 95 % confidence interval for the estimated regression, computed via bootstrap.
Fig. 4
Fig. 4
Examples of the segmentations and the correspondent score maps for a cervical cancer case (left, HiS = 0.76) and a rectal cancer case (right, HiS = 0.89). The input images for the segmentation framework, the ground truth segmentation (green) and the automatic segmentation (pink) are depicted on the top row. The corresponding score maps are depicted on the bottom row. The blue line encompasses the voxels for which the score values are higher than λ = 0.3. (For interpretation of the references to colour in this figure legend, the reader is referred to the web version of this article.)
Fig. 5
Fig. 5
AUC for detecting segmentations exceeding a specified MSD (left) or 95th HD (right).

References

    1. Liu Z., Tong L., Chen L., Jiang Z., Zhou F., Zhang Q., et al. Deep learning based brain tumor segmentation: a survey. Complex Intell Syst. 2023;9:1001–1026. doi: 10.1007/s40747-022-00815-5. - DOI
    1. Biratu E.S., Schwenker F., Ayano Y.M., Debelee T.G. A survey of brain tumor segmentation and classification algorithms. J Imaging. 2021;7 doi: 10.3390/jimaging7090179. - DOI - PMC - PubMed
    1. Ren J., Eriksen J.G., Nijkamp J., Korreman S.S. Comparing different CT, PET and MRI multi-modality image combinations for deep learning-based head and neck tumor segmentation. Acta Oncol. 2021;60:1399–1406. doi: 10.1080/0284186X.2021.1949034. - DOI - PubMed
    1. Wahid K.A., Ahmed S., He R., van Dijk L.V., Teuwen J., McDonald B.A., et al. Evaluation of deep learning-based multiparametric MRI oropharyngeal primary tumor auto-segmentation and investigation of input channel effects: Results from a prospective imaging registry. Clin Transl Radiat Oncol. 2022;32:6–14. doi: 10.1016/j.ctro.2021.10.003. - DOI - PMC - PubMed
    1. Rodríguez Outeiral R., Bos P., Al-Mamgani A., Jasperse B., Simões R., van der Heide U.A. Oropharyngeal primary tumor segmentation for radiotherapy planning on magnetic resonance imaging using deep learning. Phys Imaging Radiat Oncol. 2021;19:39–44. doi: 10.1016/j.phro.2021.06.005. - DOI - PMC - PubMed

LinkOut - more resources