Eye Tracking for Deep Learning Segmentation Using Convolutional Neural Networks

Deep learning with convolutional neural networks (CNNs) has experienced tremendous growth in multiple healthcare applications and has been shown to have high accuracy in semantic segmentation of medical (e.g., radiology and pathology) images. However, a key barrier in the required training of CNNs is obtaining large-scale and precisely annotated imaging data. We sought to address the lack of annotated data with eye tracking technology. As a proof of principle, our hypothesis was that segmentation masks generated with the help of eye tracking (ET) would be very similar to those rendered by hand annotation (HA). Additionally, our goal was to show that a CNN trained on ET masks would be equivalent to one trained on HA masks, the latter being the current standard approach. Step 1: Screen captures of 19 publicly available radiologic images of assorted structures within various modalities were analyzed. ET and HA masks for all regions of interest (ROIs) were generated from these image datasets. Step 2: Utilizing a similar approach, ET and HA masks for 356 publicly available T1-weighted postcontrast meningioma images were generated. Three hundred six of these image + mask pairs were used to train a CNN with U-net-based architecture. The remaining 50 images were used as the independent test set. Step 1: ET and HA masks for the nonneurological images had an average Dice similarity coefficient (DSC) of 0.86 between each other. Step 2: Meningioma ET and HA masks had an average DSC of 0.85 between each other. After separate training using both approaches, the ET approach performed virtually identically to HA on the test set of 50 images. The former had an area under the curve (AUC) of 0.88, while the latter had AUC of 0.87. ET and HA predictions had trimmed mean DSCs compared to the original HA maps of 0.73 and 0.74, respectively. These trimmed DSCs between ET and HA were found to be statistically equivalent with a p value of 0.015. We have demonstrated that ET can create segmentation masks suitable for deep learning semantic segmentation. Future work will integrate ET to produce masks in a faster, more natural manner that distracts less from typical radiology clinical workflow.

[1]  H L Kundel,et al.  Visual scanning, pattern recognition and decision-making in pulmonary nodule detection. , 1978, Investigative radiology.

[2]  Laura B. Machado,et al.  Multimedia-enhanced Radiology Reports: Concept, Components, and Challenges. , 2018, Radiographics : a review publication of the Radiological Society of North America, Inc.

[3]  Trafton Drew,et al.  Quantifying the costs of interruption during diagnostic radiology interpretation using mobile eye-tracking glasses , 2018, Journal of medical imaging.

[4]  H L Kundel,et al.  Searching for bone fractures: a comparison with pulmonary nodule search. , 1994, Academic radiology.

[5]  H L Kundel,et al.  Nature of expertise in searching mammograms for breast masses , 1996, Medical Imaging.

[6]  Claudia Mello-Thoms,et al.  Modeling visual search behavior of breast radiologists using a deep convolution neural network , 2018, Journal of medical imaging.

[7]  João Batista Neto,et al.  An empirical study on the effects of different types of noise in image classification tasks , 2016, ArXiv.

[8]  Berkman Sahiner,et al.  Deep learning in medical imaging and radiation therapy. , 2018, Medical physics.

[9]  Richard D. White,et al.  Automated Critical Test Findings Identification and Online Notification System Using Artificial Intelligence in Imaging. , 2017, Radiology.

[10]  G. Ripandelli,et al.  Optical coherence tomography. , 1998, Seminars in ophthalmology.

[11]  Kenji Suzuki,et al.  Overview of deep learning in medical imaging , 2017, Radiological Physics and Technology.

[12]  D. Speegle,et al.  Visual Fixation and Scan Patterns of Dentists Viewing Dental Periapical Radiographs: An Eye Tracking Pilot Study , 2018, Journal of endodontics.

[13]  D G Altman,et al.  Towards a framework for analysis of eye-tracking studies in the three dimensional environment: a study of visual search by experienced readers of endoluminal CT colonography. , 2014, The British journal of radiology.

[14]  E. Krupinski,et al.  Perceptual skill, radiology expertise, and visual test performance with NINA and WALDO. , 1998, Academic radiology.

[15]  Trafton Drew,et al.  When and why might a computer-aided detection (CAD) system interfere with visual search? An eye-tracking study. , 2012, Academic radiology.

[16]  Elizabeth A Krupinski,et al.  Tired in the Reading Room: The Influence of Fatigue in Radiology. , 2017, Journal of the American College of Radiology : JACR.

[17]  E. Krupinski,et al.  Searching for lung nodules. Visual dwell indicates locations of false-positive and false-negative decisions. , 1989, Investigative radiology.

[18]  Elizabeth A. Krupinski,et al.  Research and applications: Investigating the link between radiologists' gaze, diagnostic decision, and image content , 2013, J. Am. Medical Informatics Assoc..

[19]  J. Fujimoto,et al.  Optical Coherence Tomography , 1991 .

[20]  Xinjian Chen,et al.  Gaze2Segment: A Pilot Study for Integrating Eye-Tracking Technology into Medical Image Segmentation , 2016, MCV/BAMBI@MICCAI.

[21]  Deniz Erdogmus,et al.  Auto-Context Convolutional Neural Network (Auto-Net) for Brain Extraction in Magnetic Resonance Imaging , 2017, IEEE Transactions on Medical Imaging.

[22]  Olivier Clatz,et al.  A review of existing and potential computer user interfaces for modern radiology , 2018, Insights into Imaging.

[23]  Nico Karssemeijer,et al.  Using deep learning to segment breast and fibroglandular tissue in MRI volumes , 2017, Medical physics.

[24]  E. Krupinski,et al.  Eye-movement study and human performance using telepathology virtual slides: implications for medical education and differences with experience. , 2006, Human pathology.

[25]  Raymond Bond,et al.  Computing eye gaze metrics for the automatic assessment of radiographer performance during X-ray image interpretation , 2017, Int. J. Medical Informatics.

[26]  Donald J. Schuirmann A comparison of the Two One-Sided Tests Procedure and the Power Approach for assessing the equivalence of average bioavailability , 1987, Journal of Pharmacokinetics and Biopharmaceutics.

[27]  Lina J. Karam,et al.  Understanding how image quality affects deep neural networks , 2016, 2016 Eighth International Conference on Quality of Multimedia Experience (QoMEX).

[28]  Thomas Brox,et al.  U-Net: Convolutional Networks for Biomedical Image Segmentation , 2015, MICCAI.

[29]  Joseph N Stember,et al.  Convolutional Neural Networks for the Detection and Measurement of Cerebral Aneurysms on Magnetic Resonance Angiography , 2018, Journal of Digital Imaging.

[30]  Elizabeth A Krupinski,et al.  Search pattern training for evaluation of central venous catheter positioning on chest radiographs , 2018, Journal of medical imaging.

[31]  Ulas Bagci,et al.  A collaborative computer aided diagnosis (C‐CAD) system with eye‐tracking, sparse attentional model, and deep learning☆ , 2018, Medical Image Anal..

[32]  Elizabeth A Krupinski,et al.  The Effects of Fatigue From Overnight Shifts on Radiology Search Patterns and Diagnostic Performance. , 2018, Journal of the American College of Radiology : JACR.

[33]  D. Lakens Equivalence Tests , 2017, Social psychological and personality science.

[34]  Bram van Ginneken,et al.  A survey on deep learning in medical image analysis , 2017, Medical Image Anal..