Contrastive Explanations In Neural Networks

Visual explanations are logical arguments based on visual features that justify the predictions made by neural networks. Current modes of visual explanations answer questions of the form ‘Why P.7’. These Why questions operate under broad contexts thereby providing answers that are irrelevant in some cases. We propose to constrain these Why questions based on some context Q so that our explanations answer contrastive questions of the form ‘Why P, rather than Q.7’. In this paper, we formalize the structure of contrastive visual explanations for neural networks. We define contrast based on neural networks and propose a methodology to extract defined contrasts. We then use the extracted contrasts as a plug-in on top of existing ‘Why P?’ techniques, specifically Grad-CAM. We demonstrate their value in analyzing both networks and data in applications of largescale recognition, fine-grained recognition, subsurface seismic analysis, and image quality assessment.

[1]  D. Chandler Seven Challenges in Image Quality Assessment: Past, Present, and Future Research , 2013 .

[2]  Andrew Zisserman,et al.  Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.

[3]  Sebastian Bosse,et al.  Deep Neural Networks for No-Reference and Full-Reference Image Quality Assessment , 2016, IEEE Transactions on Image Processing.

[4]  Kilian Q. Weinberger,et al.  Densely Connected Convolutional Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[5]  Michael S. Bernstein,et al.  ImageNet Large Scale Visual Recognition Challenge , 2014, International Journal of Computer Vision.

[6]  Joseph C. Pitt,et al.  Theories of Explanation , 1988 .

[7]  C. Hempel,et al.  Studies in the Logic of Explanation , 1948, Philosophy of Science.

[8]  Sophie Pfeifer,et al.  Theories Of Explanation , 2016 .

[9]  Andrew Zisserman,et al.  Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps , 2013, ICLR.

[10]  F. Keil,et al.  Explanation and understanding , 2015 .

[11]  Ghassan Al-Regib,et al.  A curvelet-based distance measure for seismic images , 2015, 2015 IEEE International Conference on Image Processing (ICIP).

[12]  J. Pearl Causal inference in statistics: An overview , 2009 .

[13]  Ghassan Al-Regib,et al.  UNIQUE: Unsupervised Image Quality Estimation , 2016, IEEE Signal Processing Letters.

[14]  Ziyan Wu,et al.  Counterfactual Visual Explanations , 2019, ICML.

[15]  Thomas Mensink,et al.  Image Classification with the Fisher Vector: Theory and Practice , 2013, International Journal of Computer Vision.

[16]  Geoffrey E. Hinton,et al.  Learning representations by back-propagating errors , 1986, Nature.

[17]  Jonathan Krause,et al.  3D Object Representations for Fine-Grained Categorization , 2013, 2013 IEEE International Conference on Computer Vision Workshops.

[18]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[19]  Ghassan Al-Regib,et al.  Traffic Signs in the Wild: Highlights from the IEEE Video and Image Processing Cup 2017 Student Competition [SP Competitions] , 2018, IEEE Signal Process. Mag..

[20]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[21]  Ghassan Al-Regib,et al.  CURE-TSR: Challenging Unreal and Real Environments for Traffic Sign Recognition , 2017, ArXiv.

[22]  Trevor Darrell,et al.  Generating Visual Explanations , 2016, ECCV.

[23]  Antti Koura,et al.  An approach to why-questions , 1988, Synthese.

[24]  Nikolay N. Ponomarenko,et al.  Image database TID2013: Peculiarities, results and perspectives , 2015, Signal Process. Image Commun..

[25]  Linda G. Shapiro,et al.  Unsupervised Template Learning for Fine-Grained Object Recognition , 2012, NIPS.

[26]  Ghassan Al-Regib,et al.  MS-UNIQUE: Multi-model and Sharpness-weighted Unsupervised Image Quality Estimation , 2017, IQSP.

[27]  Ghassan Al-Regib,et al.  Traffic Sign Detection Under Challenging Conditions: A Deeper Look into Performance Variations and Spectral Characteristics , 2019, IEEE Transactions on Intelligent Transportation Systems.

[28]  Ghassan Al-Regib,et al.  Distorted Representation Space Characterization Through Backpropagated Gradients , 2019, 2019 IEEE International Conference on Image Processing (ICIP).

[29]  Daniel A. Wilkenfeld,et al.  Functional explaining: a new approach to the philosophy of explanation , 2014, Synthese.

[30]  David Haussler,et al.  Exploiting Generative Models in Discriminative Classifiers , 1998, NIPS.

[31]  Forrest N. Iandola,et al.  SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <1MB model size , 2016, ArXiv.