暂无分享,去创建一个
Alexandros G. Dimakis | Inderjit S. Dhillon | Michael Witbrock | Pin-Yu Chen | Qi Lei | Lingfei Wu | A. Dimakis | I. Dhillon | Qi Lei | Pin-Yu Chen | M. Witbrock | Lingfei Wu
[1] Mani B. Srivastava,et al. Generating Natural Language Adversarial Examples , 2018, EMNLP.
[2] Jinfeng Yi,et al. Seq2Sick: Evaluating the Robustness of Sequence-to-Sequence Models with Adversarial Examples , 2018, AAAI.
[3] Kevin Gimpel,et al. From Paraphrase Database to Compositional Paraphrase Model and Back , 2015, Transactions of the Association for Computational Linguistics.
[4] Dawn Song,et al. Robust Physical-World Attacks on Deep Learning Models , 2017, 1707.08945.
[5] Alexander Schrijver,et al. Combinatorial optimization. Polyhedra and efficiency. , 2003 .
[6] Dejing Dou,et al. HotFlip: White-Box Adversarial Examples for NLP , 2017, ArXiv.
[7] Pedro M. Domingos,et al. Adversarial classification , 2004, KDD.
[8] Percy Liang,et al. Adversarial Examples for Evaluating Reading Comprehension Systems , 2017, EMNLP.
[9] Ben Y. Zhao,et al. Automated Crowdturfing Attacks and Defenses in Online Review Systems , 2017, CCS.
[10] Mark W. Schmidt,et al. Coordinate Descent Converges Faster with the Gauss-Southwell Rule Than Random Selection , 2015, ICML.
[11] Fabio Roli,et al. Wild Patterns: Ten Years After the Rise of Adversarial Machine Learning , 2018, CCS.
[12] Sameep Mehta,et al. Towards Crafting Text Adversarial Samples , 2017, ArXiv.
[13] Catherine Wong,et al. DANCin SEQ2SEQ: Fooling Text Classifiers with Adversarial Text Example Generation , 2017, ArXiv.
[14] Sameer Singh,et al. Generating Natural Adversarial Examples , 2017, ICLR.
[15] H. Narayanan. Submodular functions and electrical networks , 1997 .
[16] Ananthram Swami,et al. Crafting adversarial input sequences for recurrent neural networks , 2016, MILCOM 2016 - 2016 IEEE Military Communications Conference.
[17] Andrew M. Dai,et al. Adversarial Training Methods for Semi-Supervised Text Classification , 2016, ICLR.
[18] Seyed-Mohsen Moosavi-Dezfooli,et al. DeepFool: A Simple and Accurate Method to Fool Deep Neural Networks , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[19] David Vandyke,et al. Counter-fitting Word Vectors to Linguistic Constraints , 2016, NAACL.
[20] Philip Wolfe,et al. An algorithm for quadratic programming , 1956 .
[21] Matt J. Kusner,et al. From Word Embeddings To Document Distances , 2015, ICML.
[22] Ananthram Swami,et al. Distillation as a Defense to Adversarial Perturbations Against Deep Neural Networks , 2015, 2016 IEEE Symposium on Security and Privacy (SP).
[23] Wenruo Bai,et al. Deep Submodular Functions , 2017, ArXiv.
[24] Jinfeng Yi,et al. Is Robustness the Cost of Accuracy? - A Comprehensive Study on the Robustness of 18 Deep Image Classification Models , 2018, ECCV.
[25] Yonatan Belinkov,et al. Synthetic and Natural Noise Both Break Neural Machine Translation , 2017, ICLR.
[26] Yanjun Qi,et al. Black-Box Generation of Adversarial Text Sequences to Evade Deep Learning Classifiers , 2018, 2018 IEEE Security and Privacy Workshops (SPW).
[27] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[28] Yoon Kim,et al. Convolutional Neural Networks for Sentence Classification , 2014, EMNLP.
[29] Jürgen Schmidhuber,et al. Long Short-Term Memory , 1997, Neural Computation.
[30] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[31] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[32] Carlos Guestrin,et al. Semantically Equivalent Adversarial Rules for Debugging NLP models , 2018, ACL.
[33] M. L. Fisher,et al. An analysis of approximations for maximizing submodular set functions—I , 1978, Math. Program..
[34] Xirong Li,et al. Deep Text Classification Can be Fooled , 2017, IJCAI.
[35] Xiang Zhang,et al. Character-level Convolutional Networks for Text Classification , 2015, NIPS.
[36] Michael I. Jordan,et al. Greedy Attack and Gumbel Attack: Generating Adversarial Examples for Discrete Data , 2018, J. Mach. Learn. Res..
[37] Bo Li,et al. Adversarial Texts with Gradient Methods , 2018, ArXiv.
[38] Stefano Ermon,et al. Adversarial Examples for Natural Language Classification Problems , 2018 .
[39] David A. Wagner,et al. Obfuscated Gradients Give a False Sense of Security: Circumventing Defenses to Adversarial Examples , 2018, ICML.
[40] Kevin Gimpel,et al. Pushing the Limits of Paraphrastic Sentence Embeddings with Millions of Machine Translations , 2017, ArXiv.
[41] Daniel Jurafsky,et al. Understanding Neural Networks through Representation Erasure , 2016, ArXiv.
[42] Christopher Meek,et al. Adversarial learning , 2005, KDD '05.