暂无分享,去创建一个
Nicolas Flammarion | Matthias Hein | Francesco Croce | Maksym Andriushchenko | Naman D. Singh | Naman D. Singh | Nicolas Flammarion | Matthias Hein | Maksym Andriushchenko | Francesco Croce | Nicolas Flammarion
[1] Mark Lee,et al. On Physical Adversarial Patches for Object Detection , 2019, ArXiv.
[2] Dawn Xiaodong Song,et al. Practical Black-Box Attacks on Deep Neural Networks Using Efficient Query Mechanisms , 2018, ECCV.
[3] Adam M. Oberman,et al. A principled approach for generating adversarial images under non-smooth dissimilarity metrics , 2019, AISTATS.
[4] Jinfeng Yi,et al. EAD: Elastic-Net Attacks to Deep Neural Networks via Adversarial Examples , 2017, AAAI.
[5] Seyed-Mohsen Moosavi-Dezfooli,et al. DeepFool: A Simple and Accurate Method to Fool Deep Neural Networks , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[6] David A. Wagner,et al. Obfuscated Gradients Give a False Sense of Security: Circumventing Defenses to Adversarial Examples , 2018, ICML.
[7] Olivier Teytaud,et al. Yet another but more efficient black-box adversarial attack: tiling and evolution strategies , 2019, ArXiv.
[8] Greg Yang,et al. Improved Image Wasserstein Attacks and Defenses , 2020, ArXiv.
[9] Samy Bengio,et al. Adversarial examples in the physical world , 2016, ICLR.
[10] Konrad Rieck,et al. DREBIN: Effective and Explainable Detection of Android Malware in Your Pocket , 2014, NDSS.
[11] Jun Zhu,et al. Boosting Adversarial Attacks with Momentum , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[12] Peter Szolovits,et al. Is BERT Really Robust? Natural Language Attack on Text Classification and Entailment , 2019, ArXiv.
[13] Ying Tan,et al. Generating Adversarial Malware Examples for Black-Box Attacks Based on GAN , 2017, DMBD.
[14] Una-May O'Reilly,et al. There are No Bit Parts for Sign Bits in Black-Box Attacks , 2019, ArXiv.
[15] Martín Abadi,et al. Adversarial Patch , 2017, ArXiv.
[16] Luiz Eduardo Soares de Oliveira,et al. Decoupling Direction and Norm for Efficient Gradient-Based L2 Adversarial Attacks and Defenses , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[17] Fabio Roli,et al. Evasion Attacks against Machine Learning at Test Time , 2013, ECML/PKDD.
[18] Ananthram Swami,et al. The Limitations of Deep Learning in Adversarial Settings , 2015, 2016 IEEE European Symposium on Security and Privacy (EuroS&P).
[19] Baoyuan Wu,et al. Sparse Adversarial Attack via Perturbation Factorization , 2020, ECCV.
[20] Aleksander Madry,et al. Prior Convictions: Black-Box Adversarial Attacks with Bandits and Priors , 2018, ICLR.
[21] Matthias Hein,et al. Minimally distorted Adversarial Examples with a Fast Adaptive Boundary Attack , 2019, ICML.
[22] Matthias Bethge,et al. Decision-Based Adversarial Attacks: Reliable Attacks Against Black-Box Machine Learning Models , 2017, ICLR.
[23] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[24] Peter Szolovits,et al. Is BERT Really Robust? A Strong Baseline for Natural Language Attack on Text Classification and Entailment , 2020, AAAI.
[25] Alois Knoll,et al. Guessing Smart: Biased Sampling for Efficient Black-Box Adversarial Attacks , 2018, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[26] Seyed-Mohsen Moosavi-Dezfooli,et al. SparseFool: A Few Pixels Make a Big Difference , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[27] Sijia Liu,et al. On the Design of Black-Box Adversarial Examples by Leveraging Gradient-Free Optimization and Operator Splitting Method , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[28] Bernt Schiele,et al. Adversarial Training against Location-Optimized Adversarial Patches , 2020, ECCV Workshops.
[29] Mani Srivastava,et al. GenAttack: practical black-box attacks with gradient-free optimization , 2018, GECCO.
[30] Abdullah Al-Dujaili,et al. Adversarial Deep Learning for Robust Detection of Binary Encoded Malware , 2018, 2018 IEEE Security and Privacy Workshops (SPW).
[31] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[32] Soheil Feizi,et al. Perceptual Adversarial Robustness: Defense Against Unseen Threat Models , 2020, ArXiv.
[33] Matthias Hein,et al. Provable Robustness of ReLU networks via Maximization of Linear Regions , 2018, AISTATS.
[34] Nicolas Flammarion,et al. Square Attack: a query-efficient black-box adversarial attack via random search , 2020, ECCV.
[35] Tong Zhang,et al. Black-Box Adversarial Attack with Transferable Model-based Embedding , 2020, ICLR.
[36] J. Zico Kolter,et al. Wasserstein Adversarial Examples via Projected Sinkhorn Iterations , 2019, ICML.
[37] Philippe Flajolet,et al. Birthday Paradox, Coupon Collectors, Caching Algorithms and Self-Organizing Search , 1992, Discret. Appl. Math..
[38] Raman Arora,et al. Understanding Deep Neural Networks with Rectified Linear Units , 2016, Electron. Colloquium Comput. Complex..
[39] J. Zico Kolter,et al. Adversarial camera stickers: A physical camera-based attack on deep learning systems , 2019, ICML.
[40] Alan Yuille,et al. PatchAttack: A Black-box Texture-based Attack with Reinforcement Learning , 2020, ECCV.
[41] Alexei A. Efros,et al. The Unreasonable Effectiveness of Deep Features as a Perceptual Metric , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[42] Pascal Frossard,et al. Measuring the effect of nuisance variables on classifiers , 2016, BMVC.
[43] Toon Goedemé,et al. Fooling Automated Surveillance Cameras: Adversarial Patches to Attack Person Detection , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).
[44] Pedro H. O. Pinheiro,et al. Adversarial Framing for Image and Video Classification , 2018, AAAI.
[45] Matthias Bethge,et al. Foolbox v0.8.0: A Python toolbox to benchmark the robustness of machine learning models , 2017, ArXiv.
[46] Z. Zabinsky. Random Search Algorithms , 2010 .
[47] Pushmeet Kohli,et al. Adversarial Risk and the Dangers of Evaluating Against Weak Attacks , 2018, ICML.
[48] Xiaojiang Du,et al. Adversarial Samples on Android Malware Detection Systems for IoT Systems , 2019, Sensors.
[49] Jun Zhu,et al. Improving Black-box Adversarial Attacks with a Transfer-based Prior , 2019, NeurIPS.
[50] Jinfeng Yi,et al. AutoZOOM: Autoencoder-based Zeroth Order Optimization Method for Attacking Black-box Neural Networks , 2018, AAAI.
[51] Matthias Bethge,et al. Towards the first adversarially robust neural network model on MNIST , 2018, ICLR.
[52] Jack W. Stokes,et al. Attack and Defense of Dynamic Analysis-Based, Adversarial Neural Malware Classification Models , 2017 .
[53] Hassan Takabi,et al. Effectiveness of Adversarial Examples and Defenses for Malware Classification , 2019, ArXiv.
[54] Yoav Goldberg,et al. LaVAN: Localized and Visible Adversarial Noise , 2018, ICML.
[55] Patrick D. McDaniel,et al. Adversarial Perturbations Against Deep Neural Networks for Malware Classification , 2016, ArXiv.
[56] Andrew Gordon Wilson,et al. Simple Black-box Adversarial Attacks , 2019, ICML.
[57] Nina Narodytska,et al. Simple Black-Box Adversarial Attacks on Deep Neural Networks , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).
[58] Matthias Hein,et al. Sparse and Imperceivable Adversarial Attacks , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[59] Hyun Oh Song,et al. Parsimonious Black-Box Adversarial Attacks via Efficient Combinatorial Optimization , 2019, ICML.
[60] Logan Engstrom,et al. Black-box Adversarial Attacks with Limited Queries and Information , 2018, ICML.
[61] David A. Wagner,et al. Towards Evaluating the Robustness of Neural Networks , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[62] W. Brendel,et al. Foolbox: A Python toolbox to benchmark the robustness of machine learning models , 2017 .