暂无分享,去创建一个
Qiang Liu | Zhanxing Zhu | Mao Ye | Chengyue Gong | Dinghuai Zhang | Qiang Liu | Mao Ye | Zhanxing Zhu | Dinghuai Zhang | Chengyue Gong
[1] Tom Goldstein,et al. Curse of Dimensionality on Randomized Smoothing for Certifiable Robustness , 2020, ICML.
[2] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[3] Cho-Jui Hsieh,et al. Efficient Neural Network Robustness Certification with General Activation Functions , 2018, NeurIPS.
[4] Pushmeet Kohli,et al. A Framework for robustness Certification of Smoothed Classifiers using F-Divergences , 2020, ICLR.
[5] Tommi S. Jaakkola,et al. A Stratified Approach to Robustness for Randomly Smoothed Classifiers , 2019, NeurIPS 2019.
[6] Cho-Jui Hsieh,et al. Towards Robust Neural Networks via Random Self-ensemble , 2017, ECCV.
[7] Rüdiger Ehlers,et al. Formal Verification of Piece-Wise Linear Feed-Forward Neural Networks , 2017, ATVA.
[8] Greg Yang,et al. Randomized Smoothing of All Shapes and Sizes , 2020, ICML.
[9] Chih-Hong Cheng,et al. Maximum Resilience of Artificial Neural Networks , 2017, ATVA.
[10] Di He,et al. Adversarially Robust Generalization Just Requires More Unlabeled Data , 2019, ArXiv.
[11] Alexandros G. Dimakis,et al. Provable Certificates for Adversarial Examples: Fitting a Ball in the Union of Polytopes , 2019, NeurIPS.
[12] Suman Jana,et al. Certified Robustness to Adversarial Examples with Differential Privacy , 2018, 2019 IEEE Symposium on Security and Privacy (SP).
[13] J. Zico Kolter,et al. Provable defenses against adversarial examples via the convex outer adversarial polytope , 2017, ICML.
[14] David A. Wagner,et al. Obfuscated Gradients Give a False Sense of Security: Circumventing Defenses to Adversarial Examples , 2018, ICML.
[15] Harini Kannan,et al. Adversarial Logit Pairing , 2018, NIPS 2018.
[16] Thomas B. Moeslund,et al. Long-Term Occupancy Analysis Using Graph-Based Optimisation in Thermal Imagery , 2013, 2013 IEEE Conference on Computer Vision and Pattern Recognition.
[17] Clark W. Barrett,et al. Provably Minimally-Distorted Adversarial Examples , 2017 .
[18] Haichao Zhang,et al. Defense Against Adversarial Attacks Using Feature Scattering-based Adversarial Training , 2019, NeurIPS.
[19] Stephen P. Boyd,et al. Convex Optimization , 2004, Algorithms and Theory of Computation Handbook.
[20] Binghui Wang,et al. Certified Robustness for Top-k Predictions against Adversarial Perturbations via Randomized Smoothing , 2019, ICLR.
[21] Pushmeet Kohli,et al. A Dual Approach to Scalable Verification of Deep Networks , 2018, UAI.
[22] Roman Vershynin,et al. High-Dimensional Probability , 2018 .
[23] Aditi Raghunathan,et al. Semidefinite relaxations for certifying robustness to adversarial examples , 2018, NeurIPS.
[24] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[25] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[26] J. Zico Kolter,et al. Certified Adversarial Robustness via Randomized Smoothing , 2019, ICML.
[27] Li Fei-Fei,et al. ImageNet: A large-scale hierarchical image database , 2009, CVPR.
[28] Bin Dong,et al. You Only Propagate Once: Accelerating Adversarial Training via Maximal Principle , 2019, NeurIPS.
[29] James Kuelbs,et al. Some Shift Inequalities for Gaussian Measures , 1998 .
[30] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[31] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[32] Dilin Wang,et al. Improving Neural Language Modeling via Adversarial Training , 2019, ICML.
[33] Alan L. Yuille,et al. Mitigating adversarial effects through randomization , 2017, ICLR.
[34] Greg Yang,et al. Provably Robust Deep Learning via Adversarially Trained Smoothed Classifiers , 2019, NeurIPS.
[35] Ashish Tiwari,et al. Output Range Analysis for Deep Feedforward Neural Networks , 2018, NFM.