Semantic segmentation method of underwater images based on encoder-decoder architecture
- PMID: 36006956
- PMCID: PMC9409518
- DOI: 10.1371/journal.pone.0272666
Semantic segmentation method of underwater images based on encoder-decoder architecture
Abstract
With the exploration and development of marine resources, deep learning is more and more widely used in underwater image processing. However, the quality of the original underwater images is so low that traditional semantic segmentation methods obtain poor segmentation results, such as blurred target edges, insufficient segmentation accuracy, and poor regional boundary segmentation effects. To solve these problems, this paper proposes a semantic segmentation method for underwater images. Firstly, the image enhancement based on multi-spatial transformation is performed to improve the quality of the original images, which is not common in other advanced semantic segmentation methods. Then, the densely connected hybrid atrous convolution effectively expands the receptive field and slows down the speed of resolution reduction. Next, the cascaded atrous convolutional spatial pyramid pooling module integrates boundary features of different scales to enrich target details. Finally, the context information aggregation decoder fuses the features of the shallow network and the deep network to extract rich contextual information, which greatly reduces information loss. The proposed method was evaluated on RUIE, HabCam UID, and UIEBD. Compared with the state-of-the-art semantic segmentation algorithms, the proposed method has advantages in segmentation integrity, location accuracy, boundary clarity, and detail in subjective perception. On the objective data, the proposed method achieves the highest MIOU of 68.3 and OA of 79.4, and it has a low resource consumption. Besides, the ablation experiment also verifies the effectiveness of our method.
Conflict of interest statement
The authors have declared that no competing interests exist.
Figures












Similar articles
-
A multiple-channel and atrous convolution network for ultrasound image segmentation.Med Phys. 2020 Dec;47(12):6270-6285. doi: 10.1002/mp.14512. Epub 2020 Oct 18. Med Phys. 2020. PMID: 33007105
-
Deep Neural Network-Based Semantic Segmentation of Microvascular Decompression Images.Sensors (Basel). 2021 Feb 7;21(4):1167. doi: 10.3390/s21041167. Sensors (Basel). 2021. PMID: 33562275 Free PMC article.
-
Multi-Scale Deep Neural Network Based on Dilated Convolution for Spacecraft Image Segmentation.Sensors (Basel). 2022 Jun 1;22(11):4222. doi: 10.3390/s22114222. Sensors (Basel). 2022. PMID: 35684842 Free PMC article.
-
ACCPG-Net: A skin lesion segmentation network with Adaptive Channel-Context-Aware Pyramid Attention and Global Feature Fusion.Comput Biol Med. 2023 Mar;154:106580. doi: 10.1016/j.compbiomed.2023.106580. Epub 2023 Jan 25. Comput Biol Med. 2023. PMID: 36716686 Review.
-
[Application of semantic segmentation based on convolutional neural network in medical images].Sheng Wu Yi Xue Gong Cheng Xue Za Zhi. 2020 Jun 25;37(3):533-540. doi: 10.7507/1001-5515.201906067. Sheng Wu Yi Xue Gong Cheng Xue Za Zhi. 2020. PMID: 32597097 Free PMC article. Review. Chinese.
Cited by
-
Underwater Fish Segmentation Algorithm Based on Improved PSPNet Network.Sensors (Basel). 2023 Sep 25;23(19):8072. doi: 10.3390/s23198072. Sensors (Basel). 2023. PMID: 37836901 Free PMC article.
References
-
- Li C., Saeed A., Fatih P. Underwater scene prior inspired deep underwater image and video enhancement. Pattern Recog. 2020, 98, 107038.
-
- Li Y., Chen R. UDA‐Net: Densely attention network for underwater image enhancement. IET Image Process. 2021, 15, 774–785.
-
- Wei X., Yu L., Tian S., Feng P., Ning X. Underwater target detection with an attention mechanism and improved scale. Multimedia Tools App. 2021, 80, 33747–33761.
-
- Chen L.; Liu Z.; Tong L. Underwater object detection using Invert Multi-Class Adaboost with deep learning. 2020 International Joint Conference on Neural Networks (IJCNN). IEEE, 2020.
Publication types
MeSH terms
LinkOut - more resources
Full Text Sources