Real-time automatic surgical phase recognition in laparoscopic sigmoidectomy using the convolutional neural network-based deep learning approach

Automatic surgical workflow recognition is a key component for developing the context-aware computer-assisted surgery (CA-CAS) systems. However, automatic surgical phase recognition focused on colorectal surgery has not been reported. We aimed to develop a deep learning model for automatic surgical phase recognition based on laparoscopic sigmoidectomy (Lap-S) videos, which could be used for real-time phase recognition, and to clarify the accuracies of the automatic surgical phase and action recognitions using visual information. The dataset used contained 71 cases of Lap-S. The video data were divided into frame units every 1/30 s as static images. Every Lap-S video was manually divided into 11 surgical phases (Phases 0–10) and manually annotated for each surgical action on every frame. The model was generated based on the training data. Validation of the model was performed on a set of unseen test data. Convolutional neural network (CNN)-based deep learning was also used. The average surgical time was 175 min (± 43 min SD), with the individual surgical phases also showing high variations in the duration between cases. Each surgery started in the first phase (Phase 0) and ended in the last phase (Phase 10), and phase transitions occurred 14 (± 2 SD) times per procedure on an average. The accuracy of the automatic surgical phase recognition was 91.9% and those for the automatic surgical action recognition of extracorporeal action and irrigation were 89.4% and 82.5%, respectively. Moreover, this system could perform real-time automatic surgical phase recognition at 32 fps. The CNN-based deep learning approach enabled the recognition of surgical phases and actions in 71 Lap-S cases based on manually annotated data. This system could perform automatic surgical phase recognition and automatic target surgical action recognition with high accuracy. Moreover, this study showed the feasibility of real-time automatic surgical phase recognition with high frame rate.

[1]  Nassir Navab,et al.  On-line Recognition of Surgical Activity for Monitoring in the Operating Room , 2008, AAAI.

[2]  J. Friedman Greedy function approximation: A gradient boosting machine. , 2001 .

[3]  Gwénolé Quellec,et al.  Real-Time Task Recognition in Cataract Surgery Videos Using Adaptive Spatiotemporal Polynomials , 2015, IEEE Transactions on Medical Imaging.

[4]  Lei Deng,et al.  PDRLGB: precise DNA-binding residue prediction using a light gradient boosting machine , 2018, BMC Bioinformatics.

[5]  Kevin Cleary,et al.  OR 2020: the operating room of the future. , 2004, Journal of laparoendoscopic & advanced surgical techniques. Part A.

[6]  Nassir Navab,et al.  Random Forests for Phase Detection in Surgical Workflow Analysis , 2014, IPCAI.

[7]  Jenny Dankelman,et al.  Discovery of high-level tasks in the operating room , 2011, J. Biomed. Informatics.

[8]  Danilo Miskovic,et al.  Observational clinical human reliability analysis (OCHRA) for competency assessment in laparoscopic colorectal surgery at the specialist level , 2012, Surgical Endoscopy.

[9]  Germain Forestier,et al.  Unsupervised Trajectory Segmentation for Surgical Gesture Recognition in Robotic Training , 2016, IEEE Transactions on Biomedical Engineering.

[10]  Andru Putra Twinanda,et al.  EndoNet: A Deep Architecture for Recognition Tasks on Laparoscopic Videos , 2016, IEEE Transactions on Medical Imaging.

[11]  Gwénolé Quellec,et al.  A Polynomial Model of Surgical Gestures for Real-Time Retrieval of Surgery Videos , 2012, MCBR-CDS.

[12]  Gregory D. Hager,et al.  Surgical Phase Recognition: from Instrumented ORs to Hospitals Around the World , 2016 .

[13]  Nassir Navab,et al.  Modeling and Segmentation of Surgical Workflow from Laparoscopic Video , 2010, MICCAI.

[14]  Gwénolé Quellec,et al.  Real-time analysis of cataract surgery videos using statistical models , 2017, Multimedia Tools and Applications.

[15]  Pierre Jannin,et al.  Automatic data-driven real-time segmentation and recognition of surgical workflow , 2016, International Journal of Computer Assisted Radiology and Surgery.

[16]  Nassir Navab,et al.  Automatic feature generation in endoscopic images , 2008, International Journal of Computer Assisted Radiology and Surgery.

[17]  A. Park,et al.  Advanced Devices for the Operating Room of the Future , 2003, Seminars in laparoscopic surgery.

[18]  Chi-Wing Fu,et al.  SV-RCNet: Workflow Recognition From Surgical Videos Using Recurrent Convolutional Network , 2018, IEEE Transactions on Medical Imaging.

[19]  Gwénolé Quellec,et al.  Real-time multilevel sequencing of cataract surgery videos , 2016, 2016 14th International Workshop on Content-Based Multimedia Indexing (CBMI).

[20]  Klaus Schöffmann,et al.  Temporal segmentation of laparoscopic videos into surgical phases , 2016, 2016 14th International Workshop on Content-Based Multimedia Indexing (CBMI).

[21]  J. J. van den Dobbelsteen,et al.  Surgical phase modelling in minimal invasive surgery , 2018, Surgical Endoscopy.

[22]  Gwénolé Quellec,et al.  Real-time recognition of surgical tasks in eye surgery videos , 2014, Medical Image Anal..

[23]  G. Hanna,et al.  Application of objective clinical human reliability analysis (OCHRA) in assessment of technical performance in laparoscopic rectal cancer surgery , 2016, Techniques in Coloproctology.

[24]  Cezary Szmigielski,et al.  Transfer learning with deep convolutional neural network for liver steatosis assessment in ultrasound images , 2018, International Journal of Computer Assisted Radiology and Surgery.

[25]  Nassir Navab,et al.  Recovery of Surgical Workflow Without Explicit Models , 2006, MICCAI.

[26]  Pierre Jannin,et al.  Surgical process modelling: a review , 2014, International Journal of Computer Assisted Radiology and Surgery.

[27]  M. Pera,et al.  Laparoscopic colorectal surgery: Current status and implementation of the latest technological innovations. , 2016, World journal of gastroenterology.

[28]  Sergey Ioffe,et al.  Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning , 2016, AAAI.