Understanding the Prediction Process of Deep Networks by Forests

Deep networks have continually outperformed conventional methods on a variety of tasks. However, the unexplainable aspect of deep networks limits its application in the high impact yet high risk tasks, e.g., financial credit scenario, medical diagnosis, terrorism detection and so on. Most existing interpretable approaches focus on heuristic visualization without a clear meaning. To tackle this issue, we propose a general framework to track the prediction making process (or the rules) from the input to the latent representations of deep networks. Given a deep model, the proposed method chooses a group of hidden units by considering both importance and diversity. Each selected unit is then fit by a random forest model to imitate the local structure of the overall model. For a specific input data, all paths from input to the selected units are merged in an intersection and union way to generate the final rules. We experimentally demonstrate that the proposed method can extract rules in various scenarios, ranging from age prediction, scene classification to activity classification, to interpret the prediction process. We further analyze the validity of the rules and visualize the laver-wise rules.

[1]  Bolei Zhou,et al.  Network Dissection: Quantifying Interpretability of Deep Visual Representations , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[2]  Ji Feng,et al.  Deep Forest: Towards An Alternative to Deep Neural Networks , 2017, IJCAI.

[3]  Pascal Fua,et al.  SLIC Superpixels Compared to State-of-the-Art Superpixel Methods , 2012, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[4]  Andrew Zisserman,et al.  Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.

[5]  Carlos Guestrin,et al.  "Why Should I Trust You?": Explaining the Predictions of Any Classifier , 2016, ArXiv.

[6]  Graham W. Taylor,et al.  Adaptive deconvolutional networks for mid and high level feature learning , 2011, 2011 International Conference on Computer Vision.

[7]  Stephen E. Fienberg,et al.  Homeland Insecurity : Datamining , Terrorism Detection , and Confidentiality , 2004 .

[8]  Andrea Vedaldi,et al.  Salient Deconvolutional Networks , 2016, ECCV.

[9]  Abhishek Das,et al.  Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization , 2016, 2017 IEEE International Conference on Computer Vision (ICCV).

[10]  Rob Fergus,et al.  Visualizing and Understanding Convolutional Networks , 2013, ECCV.

[11]  Bernt Schiele,et al.  2D Human Pose Estimation: New Benchmark and State of the Art Analysis , 2014, 2014 IEEE Conference on Computer Vision and Pattern Recognition.

[12]  Johannes Gehrke,et al.  Intelligible Models for HealthCare: Predicting Pneumonia Risk and Hospital 30-day Readmission , 2015, KDD.

[13]  Thomas Brox,et al.  Striving for Simplicity: The All Convolutional Net , 2014, ICLR.

[14]  Pascal Vincent,et al.  Visualizing Higher-Layer Features of a Deep Network , 2009 .

[15]  Duncan R. Wood,et al.  Governing Global Banking: The Basel Committee and the Politics of Financial Globalisation , 2005 .

[16]  Zhi-Hua Zhou,et al.  NeC4.5: neural ensemble based C4.5 , 2004, IEEE Transactions on Knowledge and Data Engineering.

[17]  Andrew Zisserman,et al.  Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps , 2013, ICLR.

[18]  Alexander Binder,et al.  On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation , 2015, PloS one.

[19]  Motoaki Kawanabe,et al.  How to Explain Individual Classification Decisions , 2009, J. Mach. Learn. Res..

[20]  Bolei Zhou,et al.  Scene Parsing through ADE20K Dataset , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[21]  Zhi-Hua Zhou,et al.  Ensemble Methods: Foundations and Algorithms , 2012 .

[22]  Bolei Zhou,et al.  Learning Deep Features for Discriminative Localization , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).