暂无分享,去创建一个
[1] Cho-Jui Hsieh,et al. On the Robustness of Self-Attentive Models , 2019, ACL.
[2] Pushmeet Kohli,et al. Verification of Non-Linear Specifications for Neural Networks , 2019, ICLR.
[3] Junfeng Yang,et al. Efficient Formal Safety Analysis of Neural Networks , 2018, NeurIPS.
[4] Sameer Singh,et al. Generating Natural Adversarial Examples , 2017, ICLR.
[5] Pushmeet Kohli,et al. A Dual Approach to Scalable Verification of Deep Networks , 2018, UAI.
[6] Aditi Raghunathan,et al. Semidefinite relaxations for certifying robustness to adversarial examples , 2018, NeurIPS.
[7] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[8] Ngai Wong,et al. POPQORN: Quantifying Robustness of Recurrent Neural Networks , 2019, ICML.
[9] Timon Gehr,et al. An abstract domain for certifying neural networks , 2019, Proc. ACM Program. Lang..
[10] Junfeng Yang,et al. Formal Security Analysis of Neural Networks using Symbolic Intervals , 2018, USENIX Security Symposium.
[11] Po-Sen Huang,et al. Achieving Verified Robustness to Symbol Substitutions via Interval Bound Propagation , 2019, EMNLP/IJCNLP.
[12] Xiang Zhang,et al. Character-level Convolutional Networks for Text Classification , 2015, NIPS.
[13] Furu Wei,et al. VL-BERT: Pre-training of Generic Visual-Linguistic Representations , 2019, ICLR.
[14] Hong Cheng,et al. Semi-Supervised Graph Classification: A Hierarchical Graph Perspective , 2019, WWW.
[15] Russ Tedrake,et al. Evaluating Robustness of Neural Networks with Mixed Integer Programming , 2017, ICLR.
[16] Jean-Baptiste Jeannin,et al. Verifying Aircraft Collision Avoidance Neural Networks Through Linear Approximations of Safe Regions , 2019, ArXiv.
[17] Pushmeet Kohli,et al. Efficient Neural Network Verification with Exactness Characterization , 2019, UAI.
[18] Julian J. McAuley,et al. Self-Attentive Sequential Recommendation , 2018, 2018 IEEE International Conference on Data Mining (ICDM).
[19] Yanjun Qi,et al. Black-Box Generation of Adversarial Text Sequences to Evade Deep Learning Classifiers , 2018, 2018 IEEE Security and Privacy Workshops (SPW).
[20] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[21] Ananthram Swami,et al. Crafting adversarial input sequences for recurrent neural networks , 2016, MILCOM 2016 - 2016 IEEE Military Communications Conference.
[22] Pushmeet Kohli,et al. A Unified View of Piecewise Linear Neural Network Verification , 2017, NeurIPS.
[23] Minlie Huang,et al. Adversarial Examples with Difficult Common Words for Paraphrase Identification , 2019 .
[24] Cho-Jui Hsieh,et al. VisualBERT: A Simple and Performant Baseline for Vision and Language , 2019, ArXiv.
[25] Adnan Darwiche,et al. A Symbolic Approach to Explaining Bayesian Network Classifiers , 2018, IJCAI.
[26] Cho-Jui Hsieh,et al. A Convex Relaxation Barrier to Tight Robustness Verification of Neural Networks , 2019, NeurIPS.
[27] Suresh Jagannathan,et al. Art: Abstraction Refinement-Guided Training for Provably Correct Neural Networks , 2019, 2020 Formal Methods in Computer Aided Design (FMCAD).
[28] Matthias Hein,et al. Formal Guarantees on the Robustness of a Classifier against Adversarial Manipulation , 2017, NIPS.
[29] Yiming Yang,et al. XLNet: Generalized Autoregressive Pretraining for Language Understanding , 2019, NeurIPS.
[30] Matthew Mirman,et al. Fast and Effective Robustness Certification , 2018, NeurIPS.
[31] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[32] Percy Liang,et al. Adversarial Examples for Evaluating Reading Comprehension Systems , 2017, EMNLP.
[33] Matthew Mirman,et al. Differentiable Abstract Interpretation for Provably Robust Neural Networks , 2018, ICML.
[34] Inderjit S. Dhillon,et al. Towards Fast Computation of Certified Robustness for ReLU Networks , 2018, ICML.
[35] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[36] Minlie Huang,et al. Robustness to Modification with Shared Words in Paraphrase Identification , 2020, EMNLP.
[37] Mani B. Srivastava,et al. Generating Natural Language Adversarial Examples , 2018, EMNLP.
[38] Alessio Lomuscio,et al. Verification of RNN-Based Neural Agent-Environment Systems , 2019, AAAI.
[39] Rüdiger Ehlers,et al. Formal Verification of Piece-Wise Linear Feed-Forward Neural Networks , 2017, ATVA.
[40] Omer Levy,et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach , 2019, ArXiv.
[41] Dejing Dou,et al. HotFlip: White-Box Adversarial Examples for Text Classification , 2017, ACL.
[42] Timothy A. Mann,et al. On the Effectiveness of Interval Bound Propagation for Training Verifiably Robust Models , 2018, ArXiv.
[43] Ashish Tiwari,et al. Output Range Analysis for Deep Feedforward Neural Networks , 2018, NFM.
[44] Cho-Jui Hsieh,et al. RecurJac: An Efficient Recursive Algorithm for Bounding Jacobian Matrix of Neural Networks and Its Applications , 2018, AAAI.
[45] Jinfeng Yi,et al. Seq2Sick: Evaluating the Robustness of Sequence-to-Sequence Models with Adversarial Examples , 2018, AAAI.
[46] Xue Liu,et al. Verification of Recurrent Neural Networks Through Rule Extraction , 2018, ArXiv.
[47] Swarat Chaudhuri,et al. AI2: Safety and Robustness Certification of Neural Networks with Abstract Interpretation , 2018, 2018 IEEE Symposium on Security and Privacy (SP).
[48] J. Zico Kolter,et al. Provable defenses against adversarial examples via the convex outer adversarial polytope , 2017, ICML.
[49] Martin Rinard,et al. Correctness Verification of Neural Networks , 2019, ArXiv.
[50] Christopher Potts,et al. Recursive Deep Models for Semantic Compositionality Over a Sentiment Treebank , 2013, EMNLP.
[51] Mykel J. Kochenderfer,et al. Reluplex: An Efficient SMT Solver for Verifying Deep Neural Networks , 2017, CAV.
[52] Cho-Jui Hsieh,et al. Efficient Neural Network Robustness Certification with General Activation Functions , 2018, NeurIPS.
[53] Aditi Raghunathan,et al. Certified Robustness to Adversarial Word Substitutions , 2019, EMNLP.