Long Term Safety Area Tracking (LT‐SAT) with online failure detection and recovery for robotic minimally invasive surgery

&NA; Despite the benefits introduced by robotic systems in abdominal Minimally Invasive Surgery (MIS), major complications can still affect the outcome of the procedure, such as intra‐operative bleeding. One of the causes is attributed to accidental damages to arteries or veins by the surgical tools, and some of the possible risk factors are related to the lack of sub‐surface visibilty. Assistive tools guiding the surgical gestures to prevent these kind of injuries would represent a relevant step towards safer clinical procedures. However, it is still challenging to develop computer vision systems able to fulfill the main requirements: (i) long term robustness, (ii) adaptation to environment/object variation and (iii) real time processing. The purpose of this paper is to develop computer vision algorithms to robustly track soft tissue areas (Safety Area, SA), defined intra‐operatively by the surgeon based on the real‐time endoscopic images, or registered from a pre‐operative surgical plan. We propose a framework to combine an optical flow algorithm with a tracking‐by‐detection approach in order to be robust against failures caused by: (i) partial occlusion, (ii) total occlusion, (iii) SA out of the field of view, (iv) deformation, (v) illumination changes, (vi) abrupt camera motion, (vii), blur and (viii) smoke. A Bayesian inference‐based approach is used to detect the failure of the tracker, based on online context information. A Model Update Strategy (MUpS) is also proposed to improve the SA re‐detection after failures, taking into account the changes of appearance of the SA model due to contact with instruments or image noise. The performance of the algorithm was assessed on two datasets, representing ex‐vivo organs and in‐vivo surgical scenarios. Results show that the proposed framework, enhanced with MUpS, is capable of maintain high tracking performance for extended periods of time ( ≃ 4 min ‐ containing the aforementioned events) with high precision (0.7) and recall (0.8) values, and with a recovery time after a failure between 1 and 8 frames in the worst case.

[1]  Danail Stoyanov,et al.  Surgical Vision , 2011, Annals of Biomedical Engineering.

[2]  Danail Stoyanov,et al.  Robust surface tracking combining features, intensity and illumination compensation , 2015, International Journal of Computer Assisted Radiology and Surgery.

[3]  Danail Stoyanov,et al.  Stereoscopic Scene Flow for Robotic Assisted Minimally Invasive Surgery , 2012, MICCAI.

[4]  Tobias Ortmaier,et al.  Soft tissue motion tracking with application to tablet-based incision planning in laser surgery , 2016, International Journal of Computer Assisted Radiology and Surgery.

[5]  W. Gantert,et al.  Bleeding remains a major complication during laparoscopic surgery: analysis of the SALTS database , 2005, Langenbeck's Archives of Surgery.

[6]  Gregory D. Hager,et al.  Articulated object tracking by rendering consistent appearance parts , 2009, 2009 IEEE International Conference on Robotics and Automation.

[7]  A. Forgione In vivo microrobots for natural orifice transluminal surgery. Current status and future perspectives. , 2009, Surgical oncology.

[8]  Leo Joskowicz,et al.  Validation of a stereo camera system to quantify brain deformation due to breathing and pulsatility. , 2014, Medical physics.

[9]  King-Sun Fu,et al.  IEEE Transactions on Pattern Analysis and Machine Intelligence Publication Information , 2004, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[10]  Iliac Vein Injury Due to a Damaged Hot Shears™ Tip Cover During Robot Assisted Radical Prostatectomy , 2011, Yonsei medical journal.

[11]  Guang-Zhong Yang,et al.  Probabilistic Tracking of Affine-Invariant Anisotropic Regions , 2013, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[12]  J. Kaouk,et al.  Robot-assisted laparoscopic partial nephrectomy: step-by-step contemporary technique and surgical outcomes at a single high-volume institution. , 2012, European urology.

[13]  G LoweDavid,et al.  Distinctive Image Features from Scale-Invariant Keypoints , 2004 .

[14]  R. Sotelo,et al.  Avoiding and managing vascular injury during robotic-assisted radical prostatectomy , 2015, Therapeutic advances in urology.

[15]  Gian Luca Mariottini,et al.  A Fast and Accurate Feature-Matching Algorithm for Minimally-Invasive Endoscopic Images , 2013, IEEE Transactions on Medical Imaging.

[16]  A. Lanfranco,et al.  Robotic Surgery: A Current Perspective , 2004, Annals of surgery.

[17]  Dietrich Paulus,et al.  Object Recognition Using Hough-transform Clustering of SURF Features , 2014 .

[18]  Leonardo S. Mattos,et al.  Virtual Assistive System for Robotic Single IncisionLaparoscopic Surgery , 2014 .

[19]  Jens Hansen,et al.  Perioperative outcomes of robot-assisted radical prostatectomy compared with open radical prostatectomy: results from the nationwide inpatient sample. , 2012, European urology.

[20]  Danail Stoyanov,et al.  Vision‐based and marker‐less surgical tool detection and tracking: a review of the literature , 2017, Medical Image Anal..

[21]  Naoki Suzuki,et al.  Identification of inferior pancreaticoduodenal artery during pancreaticoduodenectomy using augmented reality‐based navigation system , 2014, Journal of hepato-biliary-pancreatic sciences.

[22]  Robert C. Bolles,et al.  Random sample consensus: a paradigm for model fitting with applications to image analysis and automated cartography , 1981, CACM.

[23]  Nima Enayati,et al.  Haptics in Robot-Assisted Surgery: Challenges and Benefits , 2016, IEEE Reviews in Biomedical Engineering.

[24]  Jeffrey A. Cadeddu,et al.  Toward Long-Term and Accurate Augmented-Reality for Monocular Endoscopic Videos , 2014, IEEE Transactions on Biomedical Engineering.

[25]  Dietrich Paulus,et al.  Object Recognition Using Hough-transform C lustering of SURF Features , 2014 .

[26]  Dana H. Ballard,et al.  Generalizing the Hough transform to detect arbitrary shapes , 1981, Pattern Recognit..

[27]  Carlo Tomasi,et al.  Good features to track , 1994, 1994 Proceedings of IEEE Conference on Computer Vision and Pattern Recognition.

[28]  Sébastien Ourselin,et al.  Combined 2D and 3D tracking of surgical instruments for minimally invasive and robotic-assisted surgery , 2016, International Journal of Computer Assisted Radiology and Surgery.

[29]  Christopher Hunt,et al.  Notes on the OpenSURF Library , 2009 .

[30]  T M Lehmann,et al.  Color line search for illuminant estimation in real-world scenes. , 2001, Journal of the Optical Society of America. A, Optics, image science, and vision.

[31]  Guang-Zhong Yang,et al.  Online tracking and retargeting with applications to optical biopsy in gastrointestinal endoscopic examinations , 2016, Medical Image Anal..

[32]  Yi Wu,et al.  Online Object Tracking: A Benchmark , 2013, 2013 IEEE Conference on Computer Vision and Pattern Recognition.

[33]  Ling Shao,et al.  Recent advances and trends in visual tracking: A review , 2011, Neurocomputing.

[34]  M. Stifelman,et al.  Robot-assisted Laparoscopic Partial Nephrectomy , 2020, JSLS : Journal of the Society of Laparoendoscopic Surgeons.

[35]  Stefan Duffner,et al.  PixelTrack: A Fast Adaptive Algorithm for Tracking Non-rigid Objects , 2013, ICCV.

[36]  Michael C. Yip,et al.  Tissue Tracking and Registration for Image-Guided Surgery , 2012, IEEE Transactions on Medical Imaging.

[37]  Guang-Zhong Yang,et al.  Context specific descriptors for tracking deforming tissue , 2012, Medical Image Anal..

[38]  J. Marescaux,et al.  Augmented reality in laparoscopic surgical oncology. , 2011, Surgical oncology.

[39]  Luc Van Gool,et al.  SURF: Speeded Up Robust Features , 2006, ECCV.

[40]  Antonio M. Lacy,et al.  Robotics in General Surgery: Update and Future Perspectives , 2015, ICRA 2015.

[41]  Nima Enayati,et al.  EnViSoRS: Enhanced Vision System for Robotic Surgery. A User-Defined Safety Volume Tracking to Minimize the Risk of Intraoperative Bleeding , 2017, Front. Robot. AI.

[42]  Sébastien Ourselin,et al.  Image Based Surgical Instrument Pose Estimation with Multi-class Labelling and Optical Flow , 2015, MICCAI.

[43]  Bernt Schiele,et al.  Detecting Surgical Tools by Modelling Local Appearance and Global Shape , 2015, IEEE Transactions on Medical Imaging.

[44]  Elena De Momi,et al.  Dense soft tissue 3D reconstruction refined with super-pixel segmentation for robotic abdominal surgery , 2016, International Journal of Computer Assisted Radiology and Surgery.