Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2024 Mar;21(1):57-67.
doi: 10.14245/ns.2347178.589. Epub 2024 Feb 1.

Whole Spine Segmentation Using Object Detection and Semantic Segmentation

Affiliations

Whole Spine Segmentation Using Object Detection and Semantic Segmentation

Raffaele Da Mutten et al. Neurospine. 2024 Mar.

Abstract

Objective: Virtual and augmented reality have enjoyed increased attention in spine surgery. Preoperative planning, pedicle screw placement, and surgical training are among the most studied use cases. Identifying osseous structures is a key aspect of navigating a 3-dimensional virtual reconstruction. To automate the otherwise time-consuming process of labeling vertebrae on each slice individually, we propose a fully automated pipeline that automates segmentation on computed tomography (CT) and which can form the basis for further virtual or augmented reality application and radiomic analysis.

Methods: Based on a large public dataset of annotated vertebral CT scans, we first trained a YOLOv8m (You-Only-Look-Once algorithm, Version 8 and size medium) to detect each vertebra individually. On the then cropped images, a 2D-U-Net was developed and externally validated on 2 different public datasets.

Results: Two hundred fourteen CT scans (cervical, thoracic, or lumbar spine) were used for model training, and 40 scans were used for external validation. Vertebra recognition achieved a mAP50 (mean average precision with Jaccard threshold of 0.5) of over 0.84, and the segmentation algorithm attained a mean Dice score of 0.75 ± 0.14 at internal, 0.77 ± 0.12 and 0.82 ± 0.14 at external validation, respectively.

Conclusion: We propose a 2-stage approach consisting of single vertebra labeling by an object detection algorithm followed by semantic segmentation. In our externally validated pilot study, we demonstrate robust performance for our object detection network in identifying individual vertebrae, as well as for our segmentation model in precisely delineating the bony structures.

Keywords: Algorithms; Artificial intelligence; Deep learning; Machine learning; Spine.

PubMed Disclaimer

Conflict of interest statement

Conflict of Interest

The authors have nothing to disclose.

Figures

Fig. 1.
Fig. 1.
An exemplary illustration of our pipeline is shown. CT slice as input is used for object detection, then cropped and a 2DU-net for segmentation is trained and evaluated. (A) Input image with manual segmentations. (B) Object detections on CT before cropping. (C) Cropped input image for U-Net. (D) Cropped input mask for U-Net. (E) Thresholded prediction of U-Net. (F) Probability map generated from U-Net. (G) Cropped Segmentation to compare U-Net performance. (H) The cropped predictions are reassembled into a full segmentation. 2D, 2-dimensional.
Fig. 2.
Fig. 2.
Precision versus confidence plots of the YOLOv8m network, the blue line depicting performance across all classes: (A) training performance on VerSe 20, (C) holdout on VerSe 20, (E) MSD T10, (G) COVID-19. Recall versus confidence curves, the blue line depicting performance across all classes: (B) training performance on VerSe 20, (D) holdout on VerSe 20, (F) MSD T10, (H) COVID-19.
Fig. 3.
Fig. 3.
Boxplots across all 4 evaluation sets: (A) Dice score, (B) Jaccard scroe, (C) 95th percentile Hausdorff distance.
Fig. 4.
Fig. 4.
Exemplary results from external validation set. (A) CT scan from VerSe 20 holdout set. (B) A with overlay of predicted mask; red signifies high probability, blue low. (C) Ground truth to A. (D) CT scan from the MSD 10 dataset. (E) D with predictions overlay. (F) Ground truth to D. (G) CT from the COVID-19 set. (H) G with prediction overlay; red signifies high probability, blue low. (I) Ground truth to G.

References

    1. McCloskey K, Turlip R, Ahmad HS, et al. Virtual and augmented reality in spine surgery: a systematic review. World Neurosurg. 2023;173:96–107. - PubMed
    1. Mishra R, Narayanan MDK, Umana GE, et al. Virtual Reality in neurosurgery: beyond neurosurgical planning. Int J Environ Res Public Health. 2022;19:1719. - PMC - PubMed
    1. Ghaednia H, Fourman MS, Lans A, et al. Augmented and virtual reality in spine surgery, current applications and future potentials. Spine J. 2021;21:1617–25. - PubMed
    1. Alzahrani Y, Boufama B. Biomedical image segmentation: a survey. SN Comput Sci. 2021;2:310.
    1. Mharib AM, Ramli AR, Mashohor S, et al. Survey on liver CT image segmentation methods. Artif Intell Rev. 2012;37:83–95.

LinkOut - more resources