暂无分享,去创建一个
[1] Deborah Silver,et al. Feature Visualization , 1994, Scientific Visualization.
[2] Volker Tresp,et al. Saliency Methods for Explaining Adversarial Attacks , 2019, ArXiv.
[3] Nic Ford,et al. Adversarial Examples Are a Natural Consequence of Test Error in Noise , 2019, ICML.
[4] Matthias Bethge,et al. Foolbox v0.8.0: A Python toolbox to benchmark the robustness of machine learning models , 2017, ArXiv.
[5] David Wagner,et al. Adversarial Examples Are Not Easily Detected: Bypassing Ten Detection Methods , 2017, AISec@CCS.
[6] Alexander Binder,et al. On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation , 2015, PloS one.
[7] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[8] Ting Wang,et al. Interpretable Deep Learning under Fire , 2018, USENIX Security Symposium.
[9] Mingyan Liu,et al. Spatially Transformed Adversarial Examples , 2018, ICLR.
[10] Lewis D. Griffin,et al. A Boundary Tilting Persepective on the Phenomenon of Adversarial Examples , 2016, ArXiv.
[11] Prateek Mittal,et al. Dimensionality Reduction as a Defense against Evasion Attacks on Machine Learning Classifiers , 2017, ArXiv.
[12] Somesh Jha,et al. Concise Explanations of Neural Networks using Adversarial Training , 2018, ICML.
[13] P. Chalasani,et al. Adversarial Learning and Explainability in Structured Datasets. , 2018, 1810.06583.
[14] David A. Wagner,et al. Towards Evaluating the Robustness of Neural Networks , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[15] Alexander Binder,et al. Explaining nonlinear classification decisions with deep Taylor decomposition , 2015, Pattern Recognit..
[16] Klaus-Robert Müller,et al. Layer-Wise Relevance Propagation: An Overview , 2019, Explainable AI.
[17] Ajmal Mian,et al. Threat of Adversarial Attacks on Deep Learning in Computer Vision: A Survey , 2018, IEEE Access.
[18] Daniel Cullina,et al. Enhancing robustness of machine learning systems via data transformations , 2017, 2018 52nd Annual Conference on Information Sciences and Systems (CISS).
[19] Abubakar Abid,et al. Interpretation of Neural Networks is Fragile , 2017, AAAI.
[20] W. Brendel,et al. Foolbox: A Python toolbox to benchmark the robustness of machine learning models , 2017 .
[21] Dumitru Erhan,et al. The (Un)reliability of saliency methods , 2017, Explainable AI.
[22] Volker Tresp,et al. Understanding Individual Decisions of CNNs via Contrastive Backpropagation , 2018, ACCV.
[23] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[24] Yanjun Qi,et al. Feature Squeezing: Detecting Adversarial Examples in Deep Neural Networks , 2017, NDSS.
[25] Chuang Gan,et al. Interpreting Adversarial Examples by Activation Promotion and Suppression , 2019, ArXiv.
[26] Hang Su,et al. Towards Interpretable Deep Neural Networks by Leveraging Adversarial Examples , 2017, ArXiv.
[27] Eduardo Valle,et al. Exploring the space of adversarial images , 2015, 2016 International Joint Conference on Neural Networks (IJCNN).
[28] Brian Kenji Iwana,et al. Explaining Convolutional Neural Networks using Softmax Gradient Layer-wise Relevance Propagation , 2019, 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW).
[29] B. Ripley,et al. Pattern Recognition , 1968, Nature.
[30] Aleksander Madry,et al. Adversarial Examples Are Not Bugs, They Are Features , 2019, NeurIPS.