暂无分享,去创建一个
[1] Wanxiang Che,et al. Generating Natural Language Adversarial Examples through Probability Weighted Word Saliency , 2019, ACL.
[2] Provable Defenses via the Convex Outer Adversarial Polytope , 2018 .
[3] Ananthram Swami,et al. Crafting adversarial input sequences for recurrent neural networks , 2016, MILCOM 2016 - 2016 IEEE Military Communications Conference.
[4] Quan Z. Sheng,et al. Adversarial Attacks on Deep Learning Models in Natural Language Processing: A Survey , 2019 .
[5] Xiang Zhang,et al. Character-level Convolutional Networks for Text Classification , 2015, NIPS.
[6] Xiaosen Wang,et al. AT-GAN: A Generative Attack Model for Adversarial Transferring on Generative Adversarial Nets , 2019, ArXiv.
[7] Haichao Zhang,et al. Defense Against Adversarial Attacks Using Feature Scattering-based Adversarial Training , 2019, NeurIPS.
[8] Moustapha Cissé,et al. Countering Adversarial Images using Input Transformations , 2018, ICLR.
[9] Xuanjing Huang,et al. Recurrent Neural Network for Text Classification with Multi-Task Learning , 2016, IJCAI.
[10] Di He,et al. Adversarially Robust Generalization Just Requires More Unlabeled Data , 2019, ArXiv.
[11] Sergio Rojas Galeano,et al. Shielding Google's language toxicity model against adversarial attacks , 2018, ArXiv.
[12] David Vandyke,et al. Counter-fitting Word Vectors to Linguistic Constraints , 2016, NAACL.
[13] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[14] Quan Z. Sheng,et al. Generating Textual Adversarial Examples for Deep Learning Models: A Survey , 2019, ArXiv.
[15] Stefano Ermon,et al. Adversarial Examples for Natural Language Classification Problems , 2018 .
[16] David A. Wagner,et al. Obfuscated Gradients Give a False Sense of Security: Circumventing Defenses to Adversarial Examples , 2018, ICML.
[17] David A. Wagner,et al. Towards Evaluating the Robustness of Neural Networks , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[18] Po-Sen Huang,et al. Are Labels Required for Improving Adversarial Robustness? , 2019, NeurIPS.
[19] Bhuwan Dhingra,et al. Combating Adversarial Misspellings with Robust Word Recognition , 2019, ACL.
[20] Dejing Dou,et al. HotFlip: White-Box Adversarial Examples for Text Classification , 2017, ACL.
[21] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[22] Thorsten Brants,et al. One billion word benchmark for measuring progress in statistical language modeling , 2013, INTERSPEECH.
[23] Geoffrey E. Hinton,et al. ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.
[24] Kun He,et al. Improving the Generalization of Adversarial Training with Domain Adaptation , 2018, ICLR.
[25] Xirong Li,et al. Deep Text Classification Can be Fooled , 2017, IJCAI.
[26] Mani B. Srivastava,et al. Generating Natural Language Adversarial Examples , 2018, EMNLP.
[27] Jeffrey Dean,et al. Efficient Estimation of Word Representations in Vector Space , 2013, ICLR.
[28] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[29] Yoon Kim,et al. Convolutional Neural Networks for Sentence Classification , 2014, EMNLP.
[30] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[31] Ludwig Schmidt,et al. Unlabeled Data Improves Adversarial Robustness , 2019, NeurIPS.
[32] Jun Zhao,et al. Recurrent Convolutional Neural Networks for Text Classification , 2015, AAAI.
[33] Aleksander Madry,et al. Adversarially Robust Generalization Requires More Data , 2018, NeurIPS.