Limited generalizability of single deep neural network for surgical instrument segmentation in different surgical environments
- PMID: 35869249
- PMCID: PMC9307578
- DOI: 10.1038/s41598-022-16923-8
Limited generalizability of single deep neural network for surgical instrument segmentation in different surgical environments
Abstract
Clarifying the generalizability of deep-learning-based surgical-instrument segmentation networks in diverse surgical environments is important in recognizing the challenges of overfitting in surgical-device development. This study comprehensively evaluated deep neural network generalizability for surgical instrument segmentation using 5238 images randomly extracted from 128 intraoperative videos. The video dataset contained 112 laparoscopic colorectal resection, 5 laparoscopic distal gastrectomy, 5 laparoscopic cholecystectomy, and 6 laparoscopic partial hepatectomy cases. Deep-learning-based surgical-instrument segmentation was performed for test sets with (1) the same conditions as the training set; (2) the same recognition target surgical instrument and surgery type but different laparoscopic recording systems; (3) the same laparoscopic recording system and surgery type but slightly different recognition target laparoscopic surgical forceps; (4) the same laparoscopic recording system and recognition target surgical instrument but different surgery types. The mean average precision and mean intersection over union for test sets 1, 2, 3, and 4 were 0.941 and 0.887, 0.866 and 0.671, 0.772 and 0.676, and 0.588 and 0.395, respectively. Therefore, the recognition accuracy decreased even under slightly different conditions. The results of this study reveal the limited generalizability of deep neural networks in the field of surgical artificial intelligence and caution against deep-learning-based biased datasets and models.Trial Registration Number: 2020-315, date of registration: October 5, 2020.
© 2022. The Author(s).
Conflict of interest statement
The authors declare no competing interests.
Figures




Similar articles
-
Development and Validation of a Model for Laparoscopic Colorectal Surgical Instrument Recognition Using Convolutional Neural Network-Based Instance Segmentation and Videos of Laparoscopic Procedures.JAMA Netw Open. 2022 Aug 1;5(8):e2226265. doi: 10.1001/jamanetworkopen.2022.26265. JAMA Netw Open. 2022. PMID: 35984660 Free PMC article.
-
[Computer-vision-based artificial intelligence for detection and recognition of instruments and organs during radical laparoscopic gastrectomy for gastric cancer: a multicenter study].Zhonghua Wei Chang Wai Ke Za Zhi. 2024 May 25;27(5):464-470. doi: 10.3760/cma.j.cn441530-20240125-00041. Zhonghua Wei Chang Wai Ke Za Zhi. 2024. PMID: 38778686 Chinese.
-
Dual-stage semantic segmentation of endoscopic surgical instruments.Med Phys. 2024 Dec;51(12):9125-9137. doi: 10.1002/mp.17397. Epub 2024 Sep 10. Med Phys. 2024. PMID: 39255375
-
Automatic Segmentation of Multiple Organs on 3D CT Images by Using Deep Learning Approaches.Adv Exp Med Biol. 2020;1213:135-147. doi: 10.1007/978-3-030-33128-3_9. Adv Exp Med Biol. 2020. PMID: 32030668 Review.
-
Machine Learning for Surgical Phase Recognition: A Systematic Review.Ann Surg. 2021 Apr 1;273(4):684-693. doi: 10.1097/SLA.0000000000004425. Ann Surg. 2021. PMID: 33201088
Cited by
-
Artificial Intelligence in Surgery and Its Potential for Gastric Cancer.J Gastric Cancer. 2023 Jul;23(3):400-409. doi: 10.5230/jgc.2023.23.e27. J Gastric Cancer. 2023. PMID: 37553128 Free PMC article. Review.
-
Use of artificial intelligence in the analysis of digital videos of invasive surgical procedures: scoping review.BJS Open. 2025 Jul 1;9(4):zraf073. doi: 10.1093/bjsopen/zraf073. BJS Open. 2025. PMID: 40673467 Free PMC article.
References
MeSH terms
LinkOut - more resources
Full Text Sources