暂无分享,去创建一个
[1] Aleksander Madry,et al. Robustness May Be at Odds with Accuracy , 2018, ICLR.
[2] Ludwig Schmidt,et al. Unlabeled Data Improves Adversarial Robustness , 2019, NeurIPS.
[3] Masashi Sugiyama,et al. Lipschitz-Margin Training: Scalable Certification of Perturbation Invariance for Deep Neural Networks , 2018, NeurIPS.
[4] Cho-Jui Hsieh,et al. Towards Stable and Efficient Training of Verifiably Robust Neural Networks , 2019, ICLR.
[5] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[6] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[7] Marti A. Hearst. Trends & Controversies: Support Vector Machines , 1998, IEEE Intell. Syst..
[8] Shin Ishii,et al. Distributional Smoothing by Virtual Adversarial Examples , 2015, ICLR.
[9] James Bailey,et al. On the Convergence and Robustness of Adversarial Training , 2021, ICML.
[10] Seyed-Mohsen Moosavi-Dezfooli,et al. Robustness via Curvature Regularization, and Vice Versa , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[11] David A. Wagner,et al. Towards Evaluating the Robustness of Neural Networks , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[12] J. Zico Kolter,et al. Overfitting in adversarially robust deep learning , 2020, ICML.
[13] Nikos Komodakis,et al. Wide Residual Networks , 2016, BMVC.
[14] Michael P. Wellman,et al. Towards the Science of Security and Privacy in Machine Learning , 2016, ArXiv.
[15] J. Zico Kolter,et al. Provable defenses against adversarial examples via the convex outer adversarial polytope , 2017, ICML.
[16] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[17] Michael I. Jordan,et al. Theoretically Principled Trade-off between Robustness and Accuracy , 2019, ICML.
[18] Dawn Xiaodong Song,et al. Curriculum Adversarial Training , 2018, IJCAI.
[19] Ruitong Huang,et al. Max-Margin Adversarial (MMA) Training: Direct Input Space Margin Maximization through Adversarial Training , 2018, ICLR.
[20] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[21] Pin-Yu Chen,et al. CAT: Customized Adversarial Training for Improved Robustness , 2020, IJCAI.
[22] Gang Niu,et al. Attacks Which Do Not Kill Training Make Adversarial Learning Stronger , 2020, ICML.
[23] Samy Bengio,et al. Understanding deep learning requires rethinking generalization , 2016, ICLR.
[24] John Duchi,et al. Understanding and Mitigating the Tradeoff Between Robustness and Accuracy , 2020, ICML.
[25] Jason Yosinski,et al. Deep neural networks are easily fooled: High confidence predictions for unrecognizable images , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[26] Cyrus Rashtchian,et al. Adversarial Robustness Through Local Lipschitzness , 2020, ArXiv.
[27] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[28] J. Zico Kolter,et al. Certified Adversarial Robustness via Randomized Smoothing , 2019, ICML.
[29] Supriyo Chakraborty,et al. Improving Adversarial Robustness Through Progressive Hardening , 2020, ArXiv.
[30] Mislav Balunovic,et al. Adversarial Training and Provable Defenses: Bridging the Gap , 2020, ICLR.
[31] Andrew Y. Ng,et al. Reading Digits in Natural Images with Unsupervised Feature Learning , 2011 .
[32] James Bailey,et al. Improving Adversarial Robustness Requires Revisiting Misclassified Examples , 2020, ICLR.