Fragile neural networks: the importance of image standardization for deep learning in digital pathology

Recently in the field of digital pathology, there have been promising advances with regards to deep learning for pathological images. These methods are often considered “black boxes”, where tracing inputs to outputs and diagnosing errors is a difficult task. This is important as neural networks are fragile, and dataset variation, which in digital pathology is attributed to biological variance, can cause low accuracy. In deep learning, this is typically addressed by adding data to the training set. However, training is costly and time-consuming to create and may not address all variation seen in these images. Digitized histology carries a great deal of variation across many dimensions (color / stain variation, lighting intensity, presentation of a disease, etc.), and some of these “low-level” image variations may cause a deep network to break due to their fragility. In this work, we use a unique dataset – cases of serially-registered H and E tissue samples from oral cavity cancer (OCC) patients – to explore the errors of a classifier trained to identify and segment different tissue types. Registered serial sections allow us to eliminate variability due to biological structure and focus on image variability including staining and lighting, and try to identify sources of error that may cause deep learning to fail. We find that perceptually-insignificant changes in an image (minor lighting and color shifts) can result in extremely poor classification performance, even when the training process tries to prevent overfitting. This suggests that great care must be taken to augment and normalize datasets to prevent errors.

[1]  A. Chattopadhyay,et al.  Oral cavity and oropharyngeal cancer incidence trends and disparities in the United States: 2000-2010. , 2015, Cancer epidemiology.

[2]  Fei-Fei Li,et al.  ImageNet: A large-scale hierarchical image database , 2009, 2009 IEEE Conference on Computer Vision and Pattern Recognition.

[3]  Andrew Y. Ng,et al.  Reading Digits in Natural Images with Unsupervised Feature Learning , 2011 .

[4]  Scott Doyle,et al.  Active deep learning: Improved training efficiency of convolutional neural networks for tissue classification in oral cavity cancer , 2018, 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018).

[5]  Scott Doyle,et al.  Registration parameter optimization for 3D tissue modeling from resected tumors cut into serial H and E slides , 2018, Medical Imaging.

[6]  A. Jemal,et al.  Cancer statistics, 2017 , 2017, CA: a cancer journal for clinicians.

[7]  Fei-Yue Wang,et al.  Traffic Flow Prediction With Big Data: A Deep Learning Approach , 2015, IEEE Transactions on Intelligent Transportation Systems.

[8]  D. Sessions,et al.  Analysis of Treatment Results for Oral Tongue Cancer , 2002, The Laryngoscope.