Layer-Wise Relevance Propagation: An Overview
暂无分享,去创建一个
Klaus-Robert Müller | Wojciech Samek | Sebastian Lapuschkin | Grégoire Montavon | Alexander Binder | K. Müller | G. Montavon | S. Lapuschkin | W. Samek | A. Binder | Klaus Müller | Sebastian Lapuschkin | Alexander Binder
[1] Klaus-Robert Müller,et al. Understanding Patch-Based Learning of Video Data by Explaining Predictions , 2019, Explainable AI.
[2] Klaus-Robert Müller,et al. Machine learning of accurate energy-conserving molecular force fields , 2016, Science Advances.
[3] Alexander Binder,et al. Evaluating the Visualization of What a Deep Neural Network Has Learned , 2015, IEEE Transactions on Neural Networks and Learning Systems.
[4] Johanna D. Moore,et al. Explanation in second generation expert systems , 1993 .
[5] P. Baldi,et al. Searching for exotic particles in high-energy physics with deep learning , 2014, Nature Communications.
[6] Klaus-Robert Müller,et al. Structuring Neural Networks for More Explainable Predictions , 2018 .
[7] Motoaki Kawanabe,et al. How to Explain Individual Classification Decisions , 2009, J. Mach. Learn. Res..
[8] Klaus-Robert Müller,et al. Interpretable deep neural networks for single-trial EEG classification , 2016, Journal of Neuroscience Methods.
[9] Sebastian Thrun,et al. Dermatologist-level classification of skin cancer with deep neural networks , 2017, Nature.
[10] Alexander Binder,et al. On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation , 2015, PloS one.
[11] Anna Shcherbina,et al. Not Just a Black Box: Learning Important Features Through Propagating Activation Differences , 2016, ArXiv.
[12] Emmanuel Vincent,et al. CRNN-Based Multiple DoA Estimation Using Acoustic Intensity Features for Ambisonics Recordings , 2019, IEEE Journal of Selected Topics in Signal Processing.
[13] Andrea Vedaldi,et al. Interpretable Explanations of Black Boxes by Meaningful Perturbation , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[14] Been Kim,et al. Considerations for Evaluation and Generalization in Interpretable Machine Learning , 2018 .
[15] Sergey Ioffe,et al. Rethinking the Inception Architecture for Computer Vision , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[16] Geoffrey E. Hinton,et al. Learning to Represent Spatial Transformations with Factored Higher-Order Boltzmann Machines , 2010, Neural Computation.
[17] Klaus-Robert Müller,et al. iNNvestigate neural networks! , 2018, J. Mach. Learn. Res..
[18] Alexander Binder,et al. Analyzing Classifiers: Fisher Vectors and Deep Neural Networks , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[19] Wojciech Samek,et al. Methods for interpreting and understanding deep neural networks , 2017, Digit. Signal Process..
[20] Shane Legg,et al. Human-level control through deep reinforcement learning , 2015, Nature.
[21] Alexander Binder,et al. Understanding and Comparing Deep Neural Networks for Age and Gender Classification , 2017, 2017 IEEE International Conference on Computer Vision Workshops (ICCVW).
[22] Martin Wattenberg,et al. SmoothGrad: removing noise by adding noise , 2017, ArXiv.
[23] Alexander Binder,et al. Unmasking Clever Hans predictors and assessing what machines really learn , 2019, Nature Communications.
[24] Max Welling,et al. Visualizing Deep Neural Network Decisions: Prediction Difference Analysis , 2017, ICLR.
[25] Klaus-Robert Müller,et al. Explaining Recurrent Neural Network Predictions in Sentiment Analysis , 2017, WASSA@EMNLP.
[26] Melanie Mitchell,et al. Interpreting individual classifications of hierarchical networks , 2013, 2013 IEEE Symposium on Computational Intelligence and Data Mining (CIDM).
[27] Günter Klambauer,et al. DeepTox: Toxicity Prediction using Deep Learning , 2016, Front. Environ. Sci..
[28] Cristian S. Calude,et al. The Deluge of Spurious Correlations in Big Data , 2016, Foundations of Science.
[29] Rob Fergus,et al. Visualizing and Understanding Convolutional Networks , 2013, ECCV.
[30] Stephen Bazen,et al. The Taylor Decomposition: A Unified Generalization of the Oaxaca Method to Nonlinear Models , 2013 .
[31] Klaus-Robert Müller,et al. Towards Explaining Anomalies: A Deep Taylor Decomposition of One-Class Models , 2018, Pattern Recognit..
[32] Tat-Seng Chua,et al. Neural Collaborative Filtering , 2017, WWW.
[33] Stan Matwin,et al. Using Qualitative Models to Guide Inductive Learning , 1993, ICML.
[34] Alexander Binder,et al. Explaining nonlinear classification decisions with deep Taylor decomposition , 2015, Pattern Recognit..
[35] Zhe L. Lin,et al. Top-Down Neural Attention by Excitation Backprop , 2016, International Journal of Computer Vision.
[36] Klaus-Robert Müller,et al. Identifying Individual Facial Expressions by Deconstructing a Neural Network , 2016, GCPR.
[37] Masaru Ishii,et al. Towards computational fluorescence microscopy: Machine learning-based integrated prediction of morphological and molecular tumor profiles , 2018, ArXiv.
[38] Andrew Zisserman,et al. Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.
[39] Ankur Taly,et al. Axiomatic Attribution for Deep Networks , 2017, ICML.
[40] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[41] Hinrich Schütze,et al. Evaluating neural network explanation methods using hybrid documents and morphosyntactic agreement , 2018, ACL.
[42] Pan He,et al. Adversarial Examples: Attacks and Defenses for Deep Learning , 2017, IEEE Transactions on Neural Networks and Learning Systems.
[43] Tim Güneysu,et al. Deep Neural Network Attribution Methods for Leakage Analysis and Symmetric Key Recovery , 2019, IACR Cryptol. ePrint Arch..
[44] David Ryan Koes,et al. Visualizing Convolutional Neural Network Protein-Ligand Scoring , 2018, Journal of molecular graphics & modelling.
[45] Isabelle Guyon,et al. An Introduction to Variable and Feature Selection , 2003, J. Mach. Learn. Res..
[46] Chong Wang,et al. Deep Speech 2 : End-to-End Speech Recognition in English and Mandarin , 2015, ICML.
[47] Volker Tresp,et al. Explaining Therapy Predictions with Layer-Wise Relevance Propagation in Neural Networks , 2018, 2018 IEEE International Conference on Healthcare Informatics (ICHI).
[48] Alexandre Tkatchenko,et al. Quantum-chemical insights from deep tensor neural networks , 2016, Nature Communications.
[49] Avanti Shrikumar,et al. Learning Important Features Through Propagating Activation Differences , 2017, ICML.
[50] Srikanth Kandula,et al. Resource Management with Deep Reinforcement Learning , 2016, HotNets.
[51] Klaus-Robert Müller,et al. "What is relevant in a text document?": An interpretable machine learning approach , 2016, PloS one.
[52] Carlos Guestrin,et al. "Why Should I Trust You?": Explaining the Predictions of Any Classifier , 2016, ArXiv.
[53] Klaus-Robert Müller,et al. Explaining the unique nature of individual gait patterns with deep learning , 2018, Scientific Reports.
[54] Klaus-Robert Müller,et al. From Clustering to Cluster Explanations via Neural Networks , 2019, IEEE transactions on neural networks and learning systems.
[55] Emily Chen,et al. How do Humans Understand Explanations from Machine Learning Systems? An Evaluation of the Human-Interpretability of Explanation , 2018, ArXiv.
[56] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[57] Brian McWilliams,et al. The Shattered Gradients Problem: If resnets are the answer, then what is the question? , 2017, ICML.
[58] Shujian Huang,et al. Deep Matrix Factorization Models for Recommender Systems , 2017, IJCAI.
[59] Bernhard Schölkopf,et al. Support Vector Method for Novelty Detection , 1999, NIPS.