Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2021 Mar 29;21(7):2375.
doi: 10.3390/s21072375.

Edge-Sensitive Left Ventricle Segmentation Using Deep Reinforcement Learning

Affiliations

Edge-Sensitive Left Ventricle Segmentation Using Deep Reinforcement Learning

Jingjing Xiong et al. Sensors (Basel). .

Abstract

Deep reinforcement learning (DRL) has been utilized in numerous computer vision tasks, such as object detection, autonomous driving, etc. However, relatively few DRL methods have been proposed in the area of image segmentation, particularly in left ventricle segmentation. Reinforcement learning-based methods in earlier works often rely on learning proper thresholds to perform segmentation, and the segmentation results are inaccurate due to the sensitivity of the threshold. To tackle this problem, a novel DRL agent is designed to imitate the human process to perform LV segmentation. For this purpose, we formulate the segmentation problem as a Markov decision process and innovatively optimize it through DRL. The proposed DRL agent consists of two neural networks, i.e., First-P-Net and Next-P-Net. The First-P-Net locates the initial edge point, and the Next-P-Net locates the remaining edge points successively and ultimately obtains a closed segmentation result. The experimental results show that the proposed model has outperformed the previous reinforcement learning methods and achieved comparable performances compared with deep learning baselines on two widely used LV endocardium segmentation datasets, namely Automated Cardiac Diagnosis Challenge (ACDC) 2017 dataset, and Sunnybrook 2009 dataset. Moreover, the proposed model achieves higher F-measure accuracy compared with deep learning methods when training with a very limited number of samples.

Keywords: Markov decision process; deep reinforcement learning; double deep Q-network; image segmentation; left ventricle segmentation.

PubMed Disclaimer

Conflict of interest statement

The authors declare no conflict of interest.

Figures

Figure 1
Figure 1
An example of the segmentation results of an image on Automated Cardiac Diagnosis Challenge (ACDC) 2017 training dataset across different epochs in the training process. The ground truth (GT) boundary is plotted in blue and the magenta dots are the points found by Next-P-Net. The red pentagram represents the first edge point found by First-P-Net.
Figure 2
Figure 2
The overall process of the proposed system: The First-P-Net finds the first edge point and generates a probability map of edge points positions. The Next-P-Net locates the next point based on the previous edge point and image information.
Figure 3
Figure 3
The left image defines the nskipneighborhoods centered on the red point. The green points represent the eight skip neighborhoods of the red point. The middle image shows the defined action space and the corresponding action directions. The right image gives an example of the segmentation result. The ground truth (GT) boundary is plotted in blue and the magenta dots are the points found by Next-P-Net. The red pentagram represents the initial edge point.
Figure 4
Figure 4
Three separate reward functions: difference IoU reward function, edge distance reward function and points clustering reward function.
Figure 5
Figure 5
The architecture of First-P-Net. 3 × 3 conv or 1 × 1 conv: 3 × 3 or 1 × 1 convolution layer followed by batch normalization and ReLU activation function. resblk: revised ResNet Block. 2×: 2 upsampling. 4×: 4 upsampling. 0.5×: 0.5 downsampling. 1/8×: 1/8 downsampling.
Figure 6
Figure 6
The architecture of Next-P-Net. 7 × 7 conv: 7 × 7 convolution layer followed by batch normalization and ReLU activation function. resblk: ResNet Block.
Figure 7
Figure 7
Examples of segmentation outcomes. The first three rows are the segmentation performances on Sunnybrook 2009 testing dataset and the last three rows are the segmentation performances on ACDC 2017 testing dataset. The ground truth (GT) boundary is plotted in blue and the magenta dots are the points found by Next-P-Net. The red pentagram represents the first edge point found by First-P-Net.
Figure 8
Figure 8
Examples of the first edge point found by First-P-Net on ACDC 2017 testing dataset. The red pentagram represents the first edge point and the small image on the upper left or upper right corner is the partial enlargement of the first point.
Figure 9
Figure 9
States with different Q-values on ACDC 2017 testing Dataset. The first two rows show some images centered on the yellow point with low Q-value, while the last two rows show some images centered on the yellow point with high Q-value.
Figure 10
Figure 10
The changes in three separate reward values, total reward value, F-measure accuracy and APD accuracy according to the learning iterations in the training process on ACDC 2017 Dataset.
Figure 11
Figure 11
The changes in APD, F-measure, Precision and Recall according to the learning epochs in the training process on ACDC 2017 dataset.

References

    1. Rundo L., Militello C., Vitabile S., Casarino C., Russo G., Midiri M., Gilardi M.C. Combining split-and-merge and multi-seed region growing algorithms for uterine fibroid segmentation in MRgFUS treatments. Med. Biol. Eng. Comput. 2016;54:1071–1084. doi: 10.1007/s11517-015-1404-6. - DOI - PubMed
    1. Ronneberger O., Fischer P., Brox T. U-net: Convolutional networks for biomedical image segmentation; Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention; Munich, Germany. 5–9 October 2015; pp. 234–241.
    1. Noh H., Hong S., Han B. Learning deconvolution network for semantic segmentation; Proceedings of the IEEE International Conference on Computer Vision; Santiago, Chile. 11–18 December 2015; pp. 1520–1528.
    1. He K., Gkioxari G., Dollar P., Girshick R. Mask r-cnn; Proceedings of the IEEE International Conference on Computer Vision; Honolulu, HI, USA. 21–26 July 2017; pp. 2961–2969.
    1. Zhao H., Shi J., Qi X., Wang X., Jia J. Pyramid scene parsing network; Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; Honolulu, HI, USA. 21–26 July 2017; pp. 2881–2890.

LinkOut - more resources