暂无分享,去创建一个
Yi Zhou | Xiaoqing Zheng | Xuanjing Huang | Cho-Jui Hsieh | Kai-wei Chang | Cho-Jui Hsieh | Kai-Wei Chang | Xuanjing Huang | Xiaoqing Zheng | Yi Zhou
[1] Thorsten Brants,et al. One billion word benchmark for measuring progress in statistical language modeling , 2013, INTERSPEECH.
[2] Pushmeet Kohli,et al. Training verified learners with learned verifiers , 2018, ArXiv.
[3] Cho-Jui Hsieh,et al. Evaluating and Enhancing the Robustness of Neural Network-based Dependency Parsing Models with Adversarial Examples , 2020, ACL.
[4] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[5] Dejing Dou,et al. HotFlip: White-Box Adversarial Examples for Text Classification , 2017, ACL.
[6] Xiang Zhang,et al. Character-level Convolutional Networks for Text Classification , 2015, NIPS.
[7] Wanxiang Che,et al. Generating Natural Language Adversarial Examples through Probability Weighted Word Saliency , 2019, ACL.
[8] Sebastian Ruder,et al. Universal Language Model Fine-tuning for Text Classification , 2018, ACL.
[9] Ananthram Swami,et al. Distillation as a Defense to Adversarial Perturbations Against Deep Neural Networks , 2015, 2016 IEEE Symposium on Security and Privacy (SP).
[10] Suman Jana,et al. Certified Robustness to Adversarial Examples with Differential Privacy , 2018, 2019 IEEE Symposium on Security and Privacy (SP).
[11] Dawn Song,et al. Robust Physical-World Attacks on Deep Learning Models , 2017, 1707.08945.
[12] J. Zico Kolter,et al. Scaling provable adversarial defenses , 2018, NeurIPS.
[13] David A. Wagner,et al. Towards Evaluating the Robustness of Neural Networks , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[14] Catherine Wong,et al. DANCin SEQ2SEQ: Fooling Text Classifiers with Adversarial Text Example Generation , 2017, ArXiv.
[15] Po-Sen Huang,et al. Achieving Verified Robustness to Symbol Substitutions via Interval Bound Propagation , 2019, EMNLP/IJCNLP.
[16] Seyed-Mohsen Moosavi-Dezfooli,et al. DeepFool: A Simple and Accurate Method to Fool Deep Neural Networks , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[17] Xirong Li,et al. Deep Text Classification Can be Fooled , 2017, IJCAI.
[18] Christopher Potts,et al. A large annotated corpus for learning natural language inference , 2015, EMNLP.
[19] Sameer Singh,et al. Generating Natural Adversarial Examples , 2017, ICLR.
[20] Cho-Jui Hsieh,et al. Evaluating and Enhancing the Robustness of Dialogue Systems: A Case Study on a Negotiation Agent , 2019, NAACL.
[21] David Vandyke,et al. Counter-fitting Word Vectors to Linguistic Constraints , 2016, NAACL.
[22] Graham Neubig,et al. On Evaluation of Adversarial Perturbations for Sequence-to-Sequence Models , 2019, NAACL.
[23] Alan L. Yuille,et al. Mitigating adversarial effects through randomization , 2017, ICLR.
[24] Zhichao Yang,et al. Word Embedding Perturbation for Sentence Classification , 2018, ArXiv.
[25] Abhimanyu Dubey,et al. Defense Against Adversarial Images Using Web-Scale Nearest-Neighbor Search , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[26] Jakob Uszkoreit,et al. A Decomposable Attention Model for Natural Language Inference , 2016, EMNLP.
[27] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[28] Tom Goldstein,et al. FreeLB: Enhanced Adversarial Training for Language Understanding , 2019, ICLR 2020.
[29] Timothy A. Mann,et al. On the Effectiveness of Interval Bound Propagation for Training Verifiably Robust Models , 2018, ArXiv.
[30] Jinfeng Yi,et al. Seq2Sick: Evaluating the Robustness of Sequence-to-Sequence Models with Adversarial Examples , 2018, AAAI.
[31] Christopher Potts,et al. Learning Word Vectors for Sentiment Analysis , 2011, ACL.
[32] Aditi Raghunathan,et al. Certified Robustness to Adversarial Word Substitutions , 2019, EMNLP.
[33] Minlie Huang,et al. Robustness Verification for Transformers , 2020, ICLR.
[34] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[35] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[36] J. Zico Kolter,et al. Certified Adversarial Robustness via Randomized Smoothing , 2019, ICML.
[37] Percy Liang,et al. Adversarial Examples for Evaluating Reading Comprehension Systems , 2017, EMNLP.
[38] Cho-Jui Hsieh,et al. Towards Stable and Efficient Training of Verifiably Robust Neural Networks , 2019, ICLR.
[39] Soheil Feizi,et al. Interpretable Adversarial Training for Text , 2019, ArXiv.
[40] Sameep Mehta,et al. Towards Crafting Text Adversarial Samples , 2017, ArXiv.
[41] Jeffrey Pennington,et al. GloVe: Global Vectors for Word Representation , 2014, EMNLP.
[42] Cho-Jui Hsieh,et al. Towards Robust Neural Networks via Random Self-ensemble , 2017, ECCV.
[43] Mani B. Srivastava,et al. Generating Natural Language Adversarial Examples , 2018, EMNLP.
[44] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[45] Andrew M. Dai,et al. Adversarial Training Methods for Semi-Supervised Text Classification , 2016, ICLR.
[46] Hiroyuki Shindo,et al. Interpretable Adversarial Perturbation in Input Embedding Space for Text , 2018, IJCAI.