Fooling Neural Network Interpretations via Adversarial Model Manipulation
暂无分享,去创建一个
[1] Franco Turini,et al. A Survey of Methods for Explaining Black Box Models , 2018, ACM Comput. Surv..
[2] Benny Pinkas,et al. Turning Your Weakness Into a Strength: Watermarking Deep Neural Networks by Backdooring , 2018, USENIX Security Symposium.
[3] Alexander Binder,et al. On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation , 2015, PloS one.
[4] Been Kim,et al. Towards A Rigorous Science of Interpretable Machine Learning , 2017, 1702.08608.
[5] Larry S. Davis,et al. Adversarial Training for Free! , 2019, NeurIPS.
[6] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[7] Luca Antiga,et al. Automatic differentiation in PyTorch , 2017 .
[8] Martin Wattenberg,et al. SmoothGrad: removing noise by adding noise , 2017, ArXiv.
[9] Rob Fergus,et al. Visualizing and Understanding Convolutional Networks , 2013, ECCV.
[10] Ting Wang,et al. Interpretable Deep Learning under Fire , 2018, USENIX Security Symposium.
[11] Scott Lundberg,et al. A Unified Approach to Interpreting Model Predictions , 2017, NIPS.
[12] Klaus-Robert Müller,et al. Investigating the influence of noise and distractors on the interpretation of neural networks , 2016, ArXiv.
[13] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[14] Stability , 1973 .
[15] Tommi S. Jaakkola,et al. Towards Robust Interpretability with Self-Explaining Neural Networks , 2018, NeurIPS.
[16] Carlos Guestrin,et al. "Why Should I Trust You?": Explaining the Predictions of Any Classifier , 2016, ArXiv.
[17] Alan L. Yuille,et al. Mitigating adversarial effects through randomization , 2017, ICLR.
[18] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[19] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[20] Bram van Ginneken,et al. A survey on deep learning in medical image analysis , 2017, Medical Image Anal..
[21] William L. Oliver,et al. The Emergence of Machine Learning Techniques in Criminology , 2013 .
[22] Ramprasaath R. Selvaraju,et al. Grad-CAM: Why did you say that? Visual Explanations from Deep Networks via Gradient-based Localization , 2016 .
[23] Toniann Pitassi,et al. Fairness through awareness , 2011, ITCS '12.
[24] P. Sedgwick. Spearman’s rank correlation coefficient , 2018, British Medical Journal.
[25] Abubakar Abid,et al. Interpretation of Neural Networks is Fragile , 2017, AAAI.
[26] Rama Chellappa,et al. Defense-GAN: Protecting Classifiers Against Adversarial Attacks Using Generative Models , 2018, ICLR.
[27] Chandan Singh,et al. Definitions, methods, and applications in interpretable machine learning , 2019, Proceedings of the National Academy of Sciences.
[28] Kate Saenko,et al. RISE: Randomized Input Sampling for Explanation of Black-box Models , 2018, BMVC.
[29] David A. Wagner,et al. Obfuscated Gradients Give a False Sense of Security: Circumventing Defenses to Adversarial Examples , 2018, ICML.
[30] Brendan Dolan-Gavitt,et al. BadNets: Identifying Vulnerabilities in the Machine Learning Model Supply Chain , 2017, ArXiv.
[31] Klaus-Robert Müller,et al. Explanations can be manipulated and geometry is to blame , 2019, NeurIPS.
[32] Samy Bengio,et al. Adversarial examples in the physical world , 2016, ICLR.
[33] Andrew Zisserman,et al. Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps , 2013, ICLR.
[34] Andrew Zisserman,et al. Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.
[35] Thomas Hofmann,et al. The Odds are Odd: A Statistical Test for Detecting Adversarial Examples , 2019, ICML.
[36] Ankur Taly,et al. Axiomatic Attribution for Deep Networks , 2017, ICML.
[37] O. Stegle,et al. Deep learning for computational biology , 2016, Molecular systems biology.
[38] Kilian Q. Weinberger,et al. Densely Connected Convolutional Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[39] Michael S. Bernstein,et al. ImageNet Large Scale Visual Recognition Challenge , 2014, International Journal of Computer Vision.
[40] Seth Flaxman,et al. European Union Regulations on Algorithmic Decision-Making and a "Right to Explanation" , 2016, AI Mag..
[41] Avanti Shrikumar,et al. Learning Important Features Through Propagating Activation Differences , 2017, ICML.
[42] Thomas Brox,et al. Striving for Simplicity: The All Convolutional Net , 2014, ICLR.
[43] Alexander Binder,et al. Evaluating the Visualization of What a Deep Neural Network Has Learned , 2015, IEEE Transactions on Neural Networks and Learning Systems.
[44] Chandan Singh,et al. Definitions, methods, and applications in interpretable machine learning , 2019, Proceedings of the National Academy of Sciences.
[45] Alexander Binder,et al. Understanding and Comparing Deep Neural Networks for Age and Gender Classification , 2017, 2017 IEEE International Conference on Computer Vision Workshops (ICCVW).
[46] Dumitru Erhan,et al. The (Un)reliability of saliency methods , 2017, Explainable AI.