Skip to main page content
U.S. flag

An official website of the United States government

Dot gov

The .gov means it’s official.
Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

Https

The site is secure.
The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

Access keys NCBI Homepage MyNCBI Homepage Main Content Main Navigation
. 2012 Apr;59(4):966-76.
doi: 10.1109/TBME.2011.2181168. Epub 2011 Dec 23.

A framework for the recognition of high-level surgical tasks from video images for cataract surgeries

Affiliations

A framework for the recognition of high-level surgical tasks from video images for cataract surgeries

F Lalys et al. IEEE Trans Biomed Eng. 2012 Apr.

Abstract

The need for a better integration of the new generation of computer-assisted-surgical systems has been recently emphasized. One necessity to achieve this objective is to retrieve data from the operating room (OR) with different sensors, then to derive models from these data. Recently, the use of videos from cameras in the OR has demonstrated its efficiency. In this paper, we propose a framework to assist in the development of systems for the automatic recognition of high-level surgical tasks using microscope videos analysis. We validated its use on cataract procedures. The idea is to combine state-of-the-art computer vision techniques with time series analysis. The first step of the framework consisted in the definition of several visual cues for extracting semantic information, therefore, characterizing each frame of the video. Five different pieces of image-based classifiers were, therefore, implemented. A step of pupil segmentation was also applied for dedicated visual cue detection. Time series classification algorithms were then applied to model time-varying data. Dynamic time warping and hidden Markov models were tested. This association combined the advantages of all methods for better understanding of the problem. The framework was finally validated through various studies. Six binary visual cues were chosen along with 12 phases to detect, obtaining accuracies of 94%.

PubMed Disclaimer

Figures

Fig. 1
Fig. 1
Framework of the recognition system.
Fig. 2
Fig. 2
Different steps of the pupil segmentation. a) input image, b) 1st step: creation of the mask, c) 2nd step: Hough transform computation, d) 3rd step: final segmentation of the pupil.
Fig. 3
Fig. 3
Different stages of the segmentation step for the detection of instruments a) the input image, b) the clean mask, c) the region of interest corresponding to the first connected component, d) the ROI corresponding to the second connected component.
Fig. 4
Fig. 4
SURF features detected on image from Fig. 3-a., and shown as blue circles. a) SURF points on the first connected component, b) SURF points on the second connected component.
Fig. 5
Fig. 5
Left-right HMM, where each state corresponds to one surgical phase
Fig. 6
Fig. 6
Typical digital microscope frames for the 12 surgical phases: 1-preparation, 2-betadine injection, 3-lateral corneal incision, 4-principal corneal incision, 5-viscoelastic injection, 6-capsulorhexis, 7-phacoemulsification, 8-cortical aspiration of the big pieces of the lens, 9- cortical aspiration of the remanescent lens, 10-expansion of the principal incision, 11-implantation of the artificial IOL, 12- adjustment of the IOL+ wound sealing
Fig. 7
Fig. 7
BVW validation studies comparison of accuracies with different number of visual words and different keypoints detectors: a) Detection of the instruments presence, b) Recognition of the cataract aspect.
Fig. 8
Fig. 8
Distance map of two surgeries and dedicated warping path using the Itakura constraint (up), along with transposition of the surgical phases (middle), and the visual cues detected by the system (down).

References

    1. Cleary K, Chung HY, Mun SK. OR 2020: The operating room of the future. Laparoendoscopic and Advanced Surgical Techniques. 2005;15(5):495–500. - PubMed
    1. Jannin P, Morandi X. Surgical models for computer-assisted neurosurgery. Neuroimage. 2007;37(3):783–91. - PubMed
    1. Bhatia B, Oates T, Xiao Y, Hu P. Real-time identification of operating room state from video. AAAI. 2007:1761–1766.
    1. Padoy N, Blum T, Feuner H, Berger MO, Navab N. On-line recognition of surgical activity for monitoring in the operating room. Proc’s of the 20th Conference on Innovative Applications of Artificial Intelligence; 2008.
    1. Speidel S, Sudra G, Senemaud J, Drentschew M, Müller-stich BP, Gun C, Dillmann R. Situation modeling and situation recognition for a context-aware augmented reality system. Progression in biomedical optics and imaging. 2008;9(1):35.