Data-driven Holistic Framework for Automated Laparoscope Optimal View Control with Learning-based Depth Perception

Laparoscopic Field of View (FOV) control is one of the most fundamental and important components in Minimally Invasive Surgery (MIS), nevertheless the traditional manual holding paradigm may easily bring fatigue to surgical assistants, and misunderstanding between surgeons also hinders assistants to provide a high-quality FOV. Targeting this problem, we here present a data-driven framework to realize an automated laparoscopic optimal FOV control. To achieve this goal, we offline learn a motion strategy of laparoscope relative to the surgeon’s hand-held surgical tool from our in-house surgical videos, developing our control domain knowledge and an optimal view generator. To adjust the laparoscope online, we first adopt a learning-based method to segment the two-dimensional (2D) position of the surgical tool, and further leverage this outcome to obtain its scale-aware depth from dense depth estimation results calculated by our novel unsupervised RoboDepth model only with the monocular camera feedback, hence in return fusing the above real-time 3D position into our control loop. To eliminate the misorientation of FOV caused by Remote Center of Motion (RCM) constraints when moving the laparoscope, we propose a novel rotation constraint using an affine map to minimize the visual warping problem, and a null-space controller is also embedded into the framework to optimize all types of errors in a unified and decoupled manner. Experiments are conducted using Universal Robot (UR) and Karl Storz Laparoscope/Instruments, which prove the feasibility of our domain knowledge and learning enabled framework for automated camera control.

[1]  Guang-Zhong Yang,et al.  Implicit gaze-assisted adaptive motion scaling for highly articulated instrument manipulation , 2017, 2017 IEEE International Conference on Robotics and Automation (ICRA).

[2]  Ankur S Agrawal,et al.  Automating endoscopic camera motion for teleoperated minimally invasive surgery using inverse reinforcement learning , 2018 .

[3]  Noah Snavely,et al.  Unsupervised Learning of Depth and Ego-Motion from Video , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[4]  J. G. Holden,et al.  Perceptual-motor coordination in an endoscopic surgery simulation , 1999, Surgical Endoscopy.

[5]  Hongdong Li,et al.  Ground-Plane-Based Absolute Scale Estimation for Monocular Visual Odometry , 2019, IEEE Transactions on Intelligent Transportation Systems.

[6]  Alois Knoll,et al.  Framework of automatic robot surgery system using Visual servoing , 2010, 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[7]  Lior Wolf,et al.  Single Image Depth Estimation Trained via Depth From Defocus Cues , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[8]  Wei Chen,et al.  Adaptive FOV Control of Laparoscopes With Programmable Composed Constraints , 2019, IEEE Transactions on Medical Robotics and Bionics.

[9]  Christophe Doignon,et al.  A Degenerate Conic-Based Method for a Direct Fitting and 3-D Pose of Cylinders with a Single Perspective View , 2007, Proceedings 2007 IEEE International Conference on Robotics and Automation.

[10]  P. Breedveld,et al.  Endoscopic camera rotation: a conceptual solution to improve hand-eye coordination in minimally-invasive surgery , 2000 .

[11]  Olaf Kähler,et al.  Object-aware bundle adjustment for correcting monocular scale drift , 2016, 2016 IEEE International Conference on Robotics and Automation (ICRA).

[12]  Guang-Zhong Yang,et al.  Real-time surgical tool tracking and pose estimation using a hybrid cylindrical marker , 2017, International Journal of Computer Assisted Radiology and Surgery.

[13]  Andreas Bihlmaier,et al.  Automated Endoscopic Camera Guidance: A Knowledge-Based System towards Robot Assisted Surgery , 2014, ISR 2014.

[14]  Danail Stoyanov,et al.  Patch-based adaptive weighting with segmentation and scale (PAWSS) for visual tracking , 2017, ArXiv.

[15]  Duncan Frost,et al.  Recovering Stable Scale in Monocular SLAM Using Object-Supplemented Bundle Adjustment , 2018, IEEE Transactions on Robotics.

[16]  P. Finlay,et al.  Controlled trial of the introduction of a robotic camera assistant (Endo Assist) for laparoscopic cholecystectomy , 2002, Surgical Endoscopy And Other Interventional Techniques.

[17]  Bernhard P. Wrobel,et al.  Multiple View Geometry in Computer Vision , 2001 .

[18]  Jonathan Kelly,et al.  Self-Supervised Scale Recovery for Monocular Depth and Egomotion Estimation , 2021, 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[19]  Henk G. Stassen,et al.  Theoretical background and conceptual solution for depth perception and eye-hand coordination problems in laparoscopic surgery , 1999 .

[20]  Guang-Zhong Yang,et al.  Gaze gesture based human robot interaction for laparoscopic surgery , 2018, Medical Image Anal..

[21]  Armin Schneider,et al.  Endoscopic Orientation Correction , 2009, MICCAI.

[22]  Oisin Mac Aodha,et al.  Unsupervised Monocular Depth Estimation with Left-Right Consistency , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[23]  Clark C. Guest,et al.  High Accuracy Monocular SFM and Scale Correction for Autonomous Driving , 2016, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[24]  Thomas Brox,et al.  U-Net: Convolutional Networks for Biomedical Image Segmentation , 2015, MICCAI.

[25]  Brady W King,et al.  Towards an autonomous robot for camera control during laparoscopic surgery. , 2013, Journal of laparoendoscopic & advanced surgical techniques. Part A.

[26]  Michel Sarkis,et al.  Multi-Task Learning for Single Image Depth Estimation and Segmentation Based on Unsupervised Network , 2020, 2020 IEEE International Conference on Robotics and Automation (ICRA).

[27]  Gabriel J. Brostow,et al.  Digging Into Self-Supervised Monocular Depth Estimation , 2018, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).