Defense Against Adversarial Images Using Web-Scale Nearest-Neighbor Search
暂无分享,去创建一个
Abhimanyu Dubey | Laurens van der Maaten | Yixuan Li | Dhruv Kumar Mahajan | Zeki Yalniz | D. Mahajan | Yixuan Li | Abhimanyu Dubey | L. Maaten | Zeki Yalniz
[1] Moustapha Cissé,et al. Countering Adversarial Images using Input Transformations , 2018, ICLR.
[2] Seyed-Mohsen Moosavi-Dezfooli,et al. DeepFool: A Simple and Accurate Method to Fool Deep Neural Networks , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[3] Thomas Brox,et al. Adversarial Examples for Semantic Image Segmentation , 2017, ICLR.
[4] Ronan Sicre,et al. Particular object retrieval with integral max-pooling of CNN activations , 2015, ICLR.
[5] Yang Song,et al. PixelDefend: Leveraging Generative Models to Understand and Defend against Adversarial Examples , 2017, ICLR.
[6] Chen Sun,et al. Revisiting Unreasonable Effectiveness of Data in Deep Learning Era , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[7] Logan Engstrom,et al. Synthesizing Robust Adversarial Examples , 2017, ICML.
[8] Luca Antiga,et al. Automatic differentiation in PyTorch , 2017 .
[9] Patrick D. McDaniel,et al. Extending Defensive Distillation , 2017, ArXiv.
[10] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[11] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[12] Rama Chellappa,et al. Defense-GAN: Protecting Classifiers Against Adversarial Attacks Using Generative Models , 2018, ICLR.
[13] Somesh Jha,et al. Analyzing the Robustness of Nearest Neighbors to Adversarial Examples , 2017, ICML.
[14] Dan Boneh,et al. Ensemble Adversarial Training: Attacks and Defenses , 2017, ICLR.
[15] Harini Kannan,et al. Adversarial Logit Pairing , 2018, NIPS 2018.
[16] David Warde-Farley,et al. 1 Adversarial Perturbations of Deep Neural Networks , 2016 .
[17] David A. Wagner,et al. Audio Adversarial Examples: Targeted Attacks on Speech-to-Text , 2018, 2018 IEEE Security and Privacy Workshops (SPW).
[18] Eric Jones,et al. SciPy: Open Source Scientific Tools for Python , 2001 .
[19] Patrick D. McDaniel,et al. Deep k-Nearest Neighbors: Towards Confident, Interpretable and Robust Deep Learning , 2018, ArXiv.
[20] Dejing Dou,et al. HotFlip: White-Box Adversarial Examples for Text Classification , 2017, ACL.
[21] Kaiming He,et al. Exploring the Limits of Weakly Supervised Pretraining , 2018, ECCV.
[22] Kyunghyun Cho,et al. Retrieval-Augmented Convolutional Neural Networks for Improved Robustness against Adversarial Examples , 2018, ArXiv.
[23] John C. Duchi,et al. Certifiable Distributional Robustness with Principled Adversarial Training , 2017, ArXiv.
[24] Ananthram Swami,et al. Practical Black-Box Attacks against Machine Learning , 2016, AsiaCCS.
[25] Zoubin Ghahramani,et al. A study of the effect of JPG compression on adversarial images , 2016, ArXiv.
[26] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[27] Martin Wattenberg,et al. Adversarial Spheres , 2018, ICLR.
[28] Jeff Johnson,et al. Billion-Scale Similarity Search with GPUs , 2017, IEEE Transactions on Big Data.
[29] Ananthram Swami,et al. Distillation as a Defense to Adversarial Perturbations Against Deep Neural Networks , 2015, 2016 IEEE Symposium on Security and Privacy (SP).
[30] Dale Schuurmans,et al. Learning with a Strong Adversary , 2015, ArXiv.
[31] Aditi Raghunathan,et al. Certified Defenses against Adversarial Examples , 2018, ICLR.
[32] David A. Shamma,et al. The New Data and New Challenges in Multimedia Research , 2015, ArXiv.
[33] Alan L. Yuille,et al. Adversarial Examples for Semantic Segmentation and Object Detection , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[34] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[35] Samy Bengio,et al. Adversarial Machine Learning at Scale , 2016, ICLR.
[36] Kibok Lee,et al. A Simple Unified Framework for Detecting Out-of-Distribution Samples and Adversarial Attacks , 2018, NeurIPS.
[37] Logan Engstrom,et al. Black-box Adversarial Attacks with Limited Queries and Information , 2018, ICML.
[38] David A. Wagner,et al. Towards Evaluating the Robustness of Neural Networks , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[39] John C. Duchi,et al. Certifying Some Distributional Robustness with Principled Adversarial Training , 2017, ICLR.