Learning 2D Surgical Camera Motion From Demonstrations

Automating camera movement during robot-assisted surgery has the potential to reduce burden on surgeons and remove the need to manually move the camera. An important sub-problem is automatic viewpoint selection, proposing camera poses that focus on important anatomical features. We use the 6 DoF Stewart Platform Research Kit (SPRK) to move the environment with a fixed endoscope, as a dual to moving the endoscope itself, to study camera motion in surgical robotics. To provide demonstrations, we link the platform's control directly to the da Vinci Research Kit (dVRK) master control system and allow control of the platform using the same pedals and tools as a clinical movable endoscope. We propose a probabilistic model that identifies image features that “dwell” close to the camera's focal point in expert demonstrations. Our experiments consider a surgical debridement scenario on silicone phantoms with inclusions of varying color and shape. We evaluate the extent to which the system correctly segments candidate debridement targets (box accuracy) and correctly ranks those targets (rank accuracy). For debridement of a single uniquely colored inclusion, the box accuracy is 80% and the rank accuracy is 100% after 100 training data points. For debridement of multiple inclusions of the same color, the box accuracy is 70.8% and the rank accuracy is 100% after 100 training data points. For debridement of inclusions of a particular shape, the box accuracy is 70.5% and the rank accuracy is 90% after 100 training data points. A demonstration video is available at: https://vimeo.com/260362958

[1]  John F. Canny,et al.  Fast and Reliable Autonomous Surgical Debridement with Cable-Driven Robots Using a Two-Phase Calibration Procedure , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).

[2]  Mateu Sbert,et al.  Viewpoint Selection using Viewpoint Entropy , 2001, VMV.

[3]  Shengyong Chen,et al.  Active vision in robotic systems: A survey of recent developments , 2011, Int. J. Robotics Res..

[4]  Septimiu E. Salcudean,et al.  A retrofit eye gaze tracker for the da Vinci and its integration in task execution using the da Vinci Research Kit , 2015, 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[5]  Josep Amat,et al.  Automatic guidance of an assistant robot in laparoscopic surgery , 1996, Proceedings of IEEE International Conference on Robotics and Automation.

[6]  Heinz Wörn,et al.  An intelligent and autonomous endoscopic guidance system for minimally invasive surgery , 2011, 2011 IEEE International Conference on Robotics and Automation.

[7]  Sanjay Krishnan,et al.  SPRK: A low-cost stewart platform for motion study in surgical robotics , 2017, 2018 International Symposium on Medical Robotics (ISMR).

[8]  Tomomasa Sato,et al.  Automatic planning of light source and camera placement for an active photometric stereo system , 1991, Proceedings. 1991 IEEE International Conference on Robotics and Automation.

[9]  Pieter Abbeel,et al.  Learning accurate kinematic control of cable-driven surgical robots using data cleaning and Gaussian Process Regression , 2014, 2014 IEEE International Conference on Automation Science and Engineering (CASE).

[10]  Michael Laskey,et al.  Using dVRK teleoperation to facilitate deep learning of automation tasks for an industrial robot , 2017, 2017 13th IEEE Conference on Automation Science and Engineering (CASE).

[11]  James A. Young,et al.  Robotic Surgical Training in an Academic Institution , 2001, Annals of surgery.

[12]  A W Partin,et al.  Comparison of robotic versus human laparoscopic camera control . , 1995, The Journal of urology.

[13]  John F. Canny,et al.  A Computational Approach to Edge Detection , 1986, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[14]  Felix Nickel,et al.  Impact of visual–spatial ability on laparoscopic camera navigation training , 2017, Surgical Endoscopy.

[15]  Colin Ware,et al.  Exploration and virtual camera control in virtual three dimensional environments , 1990, I3D '90.

[16]  David Salesin,et al.  The virtual cinematographer: a paradigm for automatic real-time camera control and directing , 1996, SIGGRAPH.

[17]  Mark R. Wilson,et al.  Psychomotor control in a virtual laparoscopic surgery training environment: gaze control parameters differentiate novices from experts , 2010, Surgical Endoscopy.

[18]  Abhilash K. Pandya,et al.  An Autonomous Camera System using the da Vinci Research Kit * , 2017 .

[19]  James Davis,et al.  Camera Placement Considering Occlusion for Robust Motion Capture , 2000 .

[20]  Ayellet Tal,et al.  Surface Regions of Interest for Viewpoint Selection , 2012, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[21]  Abhilash Pandya,et al.  A Review of Camera Viewpoint Automation in Robotic and Laparoscopic Surgery , 2014, Robotics.

[22]  A W Partin,et al.  Complete robot-assisted laparoscopic urologic surgery: a preliminary report. , 1995, Journal of the American College of Surgeons.

[23]  Christian Laugier,et al.  Automatic camera placement for robot vision tasks , 1995, Proceedings of 1995 IEEE International Conference on Robotics and Automation.

[24]  Kaiming He,et al.  Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks , 2015, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[25]  Bernhard Preim,et al.  Viewpoint Selection for Intervention Planning , 2007, EuroVis.

[26]  George P. Mylonas,et al.  Gaze-contingent control for minimally invasive robotic surgery , 2006, Computer aided surgery : official journal of the International Society for Computer Aided Surgery.

[27]  Yuichi Motai,et al.  Hand–Eye Calibration Applied to Viewpoint Selection for Robotic Vision , 2008, IEEE Transactions on Industrial Electronics.

[28]  R. Bajcsy Active perception , 1988 .

[29]  Ali Borji,et al.  State-of-the-Art in Visual Attention Modeling , 2013, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[30]  Fumihito Arai,et al.  3D viewpoint selection and bilateral control for bio-micromanipulation , 2000, Proceedings 2000 ICRA. Millennium Conference. IEEE International Conference on Robotics and Automation. Symposia Proceedings (Cat. No.00CH37065).

[31]  Patrick Olivier,et al.  Camera Control in Computer Graphics , 2008, Comput. Graph. Forum.

[32]  Joachim Denzler,et al.  Viewpoint Selection - Planning Optimal Sequences of Views for Object Recognition , 2003, CAIP.

[33]  Robert G. Moore,et al.  Laparoscopic visual field , 1998, Surgical Endoscopy.

[34]  Aaas News,et al.  Book Reviews , 1893, Buffalo Medical and Surgical Journal.

[35]  Gregory Auner,et al.  Eye Gaze Tracking for Endoscopic Camera Positioning: An Application of a Hardware/Software Interface Developed to Automate Aesop , 2008, MMVR.

[36]  J. Gilbert The EndoAssist robotic camera holder as an aid to the introduction of laparoscopic colorectal surgery. , 2009, Annals of the Royal College of Surgeons of England.