Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
Review
. 2025 May 7:34:100774.
doi: 10.1016/j.phro.2025.100774. eCollection 2025 Apr.

Uncertainties in outcome modelling in radiation oncology

Affiliations
Review

Uncertainties in outcome modelling in radiation oncology

Lukas Dünger et al. Phys Imaging Radiat Oncol. .

Abstract

Outcome models predicting e.g. survival, tumour control or radiation-induced toxicities play an important role in the field of radiation oncology. These models aim to support the clinical decision making and pave the way towards personalised treatment. Both validity and reliability of their output are required to facilitate clinical integration. However, models are influenced by uncertainties, arising from data used for model development and model parameters, among others. Therefore, quantifying model uncertainties and addressing their causes promotes the creation of models that are sufficiently reliable for clinical use. This topical review aims to summarise different types and possible sources of uncertainties, presents uncertainty quantification methods applicable to various modelling approaches, and highlights central challenges that need to be addressed in the future.

Keywords: Outcome; Radiation oncology; Statistical analyses; Uncertainty modelling.

PubMed Disclaimer

Conflict of interest statement

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper. Alex Zwanenburg is an Editorial Board Member for this journal and was not involved in the editorial review or the decision to publish this article.

Figures

None
Graphical abstract
Fig. 1
Fig. 1
Concept of data domains for addressing different types of uncertainties. (A) During the training process a mapping f(x) between the data points in the data domain xi ∈ X and the outcome domain yi ∈ Y is created, e.g. using logistic regression or a neural network. (B) When applying a model to unseen data, in-domain uncertainty refers to uncertainties for data points which are in the observed domain and have characteristics similar to the training data. In contrast, applying the model to data whose characteristics differ from the samples of the observed domain can lead to out-of-domain uncertainty. (C) In order to achieve a model that can generalise well, capturing and reducing the true domain are two main strategies for reducing uncertainties.
Fig. 2
Fig. 2
Aleatoric and epistemic uncertainty. Models 1 and 2 are trained using the same data points, but their predictions are influenced by aleatoric and epistemic uncertainties. For a classification task (A), the aleatoric uncertainty is high in the domain with overlapping data points due to inherent randomness of the data, whereas the epistemic uncertainty is high in areas with missing data points. Similarly, for a regression task (B), the epistemic uncertainty is high in areas with no or only sparse data points. In areas with numerous data points, inherent randomness results in a high aleatoric uncertainty.
Fig. 3
Fig. 3
Important elements of the data collection process. Main study designs (top, green), central questions related to data collection requirements (middle, blue) as well as relevant strategies (bottom, orange) that can be used to create a reliable dataset are shown. (For interpretation of the references to colour in this figure legend, the reader is referred to the web version of this article.)
Fig. 4
Fig. 4
Data analysis plan. A data analysis plan contains several steps that need to be addressed during the creation of a dataset for outcome modelling. The individual steps build on each other: First, in- and exclusion parameters are needed, defining the characteristics of the cohort. Second, protocols, parameters and data storage for data acquisition must be defined. Third, methods for data curation including analysis, sorting and validation of the content must be outlined. Fourth, data annotation and processing steps need to be defined, e.g. for segmentation or feature selection.
Fig. 5
Fig. 5
Methods for uncertainty quantification. Illustration of the basic principles for four different methods to quantify uncertainty using the example of a simple neural network. The methods can be used to model the relationship y = f(x) and to calculate a measure of uncertainty. Here, the variance σ2 is assessed. (A) Ensemble methods use several different networks. The prediction and a corresponding measure of uncertainty are calculated from the individual predictions. (B) Bayesian methods are based on Bayesian inference, resulting in a non-discrete model output. The resulting distribution of the outcome variable is used to calculate for instance the mean and variance. (C) Deterministic methods predict the outcome and their corresponding uncertainty simultaneously in a single-forward pass, e.g. using a connected second network in case of an external method. (D) In test-time augmentation methods, the input variable is modified resulting in several output variables. Graphical illustration after Gawlikowski et al. [18].
Fig. 6
Fig. 6
Use and representation of uncertainty quantification. Examples for applications and uncertainty representation methods are shown in the top row. These methods can be used to increase model interpretability, to address subpar model performance or to raise the need for expert referral, as illustrated in the bottom row. Abbreviations: NTCP: Normal tissue complication probability, crit.: critical value, Exp. conf.: Expected confidence, Obs. conf.: Observed confidence, fx.: fraction, Non-conf.: Non-conformity, y90%: 90% confidence prediction set.

Similar articles

References

    1. El Naqa I. 1st ed. CRC Press; 2018. A guide to outcome modeling in radiotherapy and oncology: listening to the data.
    1. Marks L.B., Yorke E.D., Jackson A., Ten Haken R.K., Constine L.S., Eisbruch A., et al. Use of normal tissue complication probability models in the clinic. Int J Radiat Oncol. 2010;76:S10–S19. doi: 10.1016/j.ijrobp.2009.07.1754. - DOI - PMC - PubMed
    1. Appelt A.L., Elhaminia B., Gooya A., Gilbert A., Nix M. Deep learning for radiotherapy outcome prediction using dose data – a review. Clinical Oncology. 2022;34:e87–e96. doi: 10.1016/j.clon.2021.12.002. - DOI - PubMed
    1. Cui S., Hope A., Dilling T.J., Dawson L.A., Ten Haken R., El Naqa I. Artificial intelligence for outcome modeling in radiotherapy. Seminars in Radiation Oncology. 2022;32:351–364. doi: 10.1016/j.semradonc.2022.06.005. - DOI - PubMed
    1. Huynh E., Hosny A., Guthier C., Bitterman D.S., Petit S.F., Haas-Kogan D.A., et al. Artificial intelligence in radiation oncology. Nature Reviews. Clinical Oncology. 2020;17:771–781. doi: 10.1038/s41571-020-0417-8. - DOI - PubMed

LinkOut - more resources