Limitations of Out-of-Distribution Detection in 3D Medical Image Segmentation
- PMID: 37754955
- PMCID: PMC10532230
- DOI: 10.3390/jimaging9090191
Limitations of Out-of-Distribution Detection in 3D Medical Image Segmentation
Abstract
Deep learning models perform unreliably when the data come from a distribution different from the training one. In critical applications such as medical imaging, out-of-distribution (OOD) detection methods help to identify such data samples, preventing erroneous predictions. In this paper, we further investigate OOD detection effectiveness when applied to 3D medical image segmentation. We designed several OOD challenges representing clinically occurring cases and found that none of the methods achieved acceptable performance. Methods not dedicated to segmentation severely failed to perform in the designed setups; the best mean false-positive rate at a 95% true-positive rate (FPR) was 0.59. Segmentation-dedicated methods still achieved suboptimal performance, with the best mean FPR being 0.31 (lower is better). To indicate this suboptimality, we developed a simple method called Intensity Histogram Features (IHF), which performed comparably or better in the same challenges, with a mean FPR of 0.25. Our findings highlight the limitations of the existing OOD detection methods with 3D medical images and present a promising avenue for improving them. To facilitate research in this area, we release the designed challenges as a publicly available benchmark and formulate practical criteria to test the generalization of OOD detection beyond the suggested benchmark. We also propose IHF as a solid baseline to contest emerging methods.
Keywords: anomaly detection; computed tomography; magnetic resonance imaging; out-of-distribution detection; segmentation.
Conflict of interest statement
The authors declare no conflict of interest.
Figures
References
-
- Wang M., Deng W. Deep Visual Domain Adaptation: A Survey. Neurocomputing. 2018;312:135–153. doi: 10.1016/j.neucom.2018.05.083. - DOI
-
- Yang J., Zhou K., Li Y., Liu Z. Generalized out-of-distribution detection: A survey. arXiv. 20212110.11334
-
- Hendrycks D., Gimpel K. A baseline for detecting misclassified and out-of-distribution examples in neural networks. arXiv. 20161610.02136
-
- Hendrycks D., Basart S., Mazeika M., Mostajabi M., Steinhardt J., Song D. Scaling out-of-distribution detection for real-world settings. arXiv. 20191911.11132
Grants and funding
LinkOut - more resources
Full Text Sources
