暂无分享,去创建一个
Geoffrey E. Hinton | Colin Raffel | Geoffrey Hinton | Yao Qin | Nicholas Frosst | Sara Sabour | Garrison Cottrell | G. Cottrell | Colin Raffel | S. Sabour | Nicholas Frosst | Yao Qin
[1] Ranga Rodrigo,et al. DeepCaps: Going Deeper With Capsule Networks , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[2] Yang Song,et al. PixelDefend: Leveraging Generative Models to Understand and Defend against Adversarial Examples , 2017, ICLR.
[3] David J. Fleet,et al. Adversarial Manipulation of Deep Representations , 2015, ICLR.
[4] Fabio Roli,et al. Evasion Attacks against Machine Learning at Test Time , 2013, ECML/PKDD.
[5] Alexei A. Efros,et al. The Unreasonable Effectiveness of Deep Features as a Perceptual Metric , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[6] Philip H. S. Torr,et al. With Friends Like These, Who Needs Adversaries? , 2018, NeurIPS.
[7] Justin Gilmer,et al. MNIST-C: A Robustness Benchmark for Computer Vision , 2019, ArXiv.
[8] Alexei A. Efros,et al. Image-to-Image Translation with Conditional Adversarial Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[9] Hao Chen,et al. MagNet: A Two-Pronged Defense against Adversarial Examples , 2017, CCS.
[10] Yao Zhao,et al. Adversarial Attacks and Defences Competition , 2018, ArXiv.
[11] Kevin Gimpel,et al. Early Methods for Detecting Adversarial Images , 2016, ICLR.
[12] Fei-Fei Li,et al. ImageNet: A large-scale hierarchical image database , 2009, 2009 IEEE Conference on Computer Vision and Pattern Recognition.
[13] Geoffrey E. Hinton,et al. Matrix capsules with EM routing , 2018, ICLR.
[14] Ryan R. Curtin,et al. Detecting Adversarial Samples from Artifacts , 2017, ArXiv.
[15] Thomas Hofmann,et al. The Odds are Odd: A Statistical Test for Detecting Adversarial Examples , 2019, ICML.
[16] David Berthelot,et al. Evaluation Methodology for Attacks Against Confidence Thresholding Models , 2018 .
[17] Matthias Bethge,et al. A note on the evaluation of generative models , 2015, ICLR.
[18] David Wagner,et al. Adversarial Examples Are Not Easily Detected: Bypassing Ten Detection Methods , 2017, AISec@CCS.
[19] Stefan Harmeling,et al. On the Vulnerability of Capsule Networks to Adversarial Attacks , 2019, ArXiv.
[20] Alexei A. Efros,et al. Unpaired Image-to-Image Translation Using Cycle-Consistent Adversarial Networks , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[21] Samy Bengio,et al. Adversarial examples in the physical world , 2016, ICLR.
[22] Taesung Park,et al. Semantic Image Synthesis With Spatially-Adaptive Normalization , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[23] David A. Wagner,et al. Obfuscated Gradients Give a False Sense of Security: Circumventing Defenses to Adversarial Examples , 2018, ICML.
[24] Andrew Y. Ng,et al. Reading Digits in Natural Images with Unsupervised Feature Learning , 2011 .
[25] Radha Poovendran,et al. Are Odds Really Odd? Bypassing Statistical Detection of Adversarial Examples , 2019, ArXiv.
[26] Ryan P. Adams,et al. Motivating the Rules of the Game for Adversarial Example Research , 2018, ArXiv.
[27] Martin Wattenberg,et al. Adversarial Spheres , 2018, ICLR.
[28] Yoshua Bengio,et al. Gradient-based learning applied to document recognition , 1998, Proc. IEEE.
[29] Krista A. Ehinger,et al. SUN database: Large-scale scene recognition from abbey to zoo , 2010, 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition.
[30] Roland Vollgraf,et al. Fashion-MNIST: a Novel Image Dataset for Benchmarking Machine Learning Algorithms , 2017, ArXiv.
[31] Geoffrey E. Hinton,et al. Dynamic Routing Between Capsules , 2017, NIPS.
[32] Rama Chellappa,et al. Defense-GAN: Protecting Classifiers Against Adversarial Attacks Using Generative Models , 2018, ICLR.
[33] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[34] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[35] Alexandros G. Dimakis,et al. The Robust Manifold Defense: Adversarial Training using Generative Models , 2017, ArXiv.
[36] Matthias Bethge,et al. Robust Perception through Analysis by Synthesis , 2018, ArXiv.
[37] Patrick D. McDaniel,et al. On the (Statistical) Detection of Adversarial Examples , 2017, ArXiv.
[38] Zhitao Gong,et al. Adversarial and Clean Data Are Not Twins , 2017, aiDM@SIGMOD.
[39] Jan Hendrik Metzen,et al. On Detecting Adversarial Perturbations , 2017, ICLR.
[40] David A. Wagner,et al. Towards Evaluating the Robustness of Neural Networks , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[41] Xin Li,et al. Adversarial Examples Detection in Deep Networks with Convolutional Filter Statistics , 2016, 2017 IEEE International Conference on Computer Vision (ICCV).
[42] Honglak Lee,et al. Learning Structured Output Representation using Deep Conditional Generative Models , 2015, NIPS.
[43] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[44] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[45] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.