暂无分享,去创建一个
Quanquan Gu | Sarah Monazam Erfani | Xingjun Ma | James Bailey | Yisen Wang | Hanxun Huang | Quanquan Gu | Yisen Wang | S. Erfani | Xingjun Ma | Hanxun Huang | James Bailey
[1] Andrew Zisserman,et al. Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.
[2] Xingyi Yang,et al. DSRNA: Differentiable Search of Robust Neural Architectures , 2020, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[3] Preetum Nakkiran,et al. Adversarial Robustness May Be at Odds With Simplicity , 2019, ArXiv.
[4] Kamyar Azizzadenesheli,et al. Stochastic Activation Pruning for Robust Adversarial Defense , 2018, ICLR.
[5] Ritu Chadha,et al. Limitations of the Lipschitz constant as a defense against adversarial examples , 2018, Nemesis/UrbReas/SoGood/IWAISe/GDM@PKDD/ECML.
[6] Yu Cheng,et al. Adversarial Robustness: From Self-Supervised Pre-Training to Fine-Tuning , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[7] Shiyu Chang,et al. Robust Overfitting may be mitigated by properly learned smoothening , 2021, ICLR.
[8] R. Venkatesh Babu,et al. Guided Adversarial Attack for Evaluating and Enhancing Adversarial Defenses , 2020, NeurIPS.
[9] Quanquan Gu,et al. Do Wider Neural Networks Really Help Adversarial Robustness? , 2020, NeurIPS.
[10] James Bailey,et al. Improving Adversarial Robustness Requires Revisiting Misclassified Examples , 2020, ICLR.
[11] Suman Jana,et al. HYDRA: Pruning Adversarially Robust Neural Networks , 2020, NeurIPS.
[12] Andrew L. Beam,et al. Adversarial attacks on medical machine learning , 2019, Science.
[13] David Doermann,et al. Anti-Bandit Neural Architecture Search for Model Defense , 2020, ECCV.
[14] Sergey Ioffe,et al. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.
[15] David A. Wagner,et al. Towards Evaluating the Robustness of Neural Networks , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[16] Kilian Q. Weinberger,et al. Densely Connected Convolutional Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[17] Di He,et al. Adversarially Robust Generalization Just Requires More Unlabeled Data , 2019, ArXiv.
[18] Jun Zhu,et al. Adversarial Distributional Training for Robust Deep Learning , 2020, NeurIPS.
[19] Micah Goldblum,et al. Adversarially Robust Distillation , 2019, AAAI.
[20] Alan L. Yuille,et al. Intriguing Properties of Adversarial Training at Scale , 2020, ICLR.
[21] Yisen Wang,et al. Adversarial Weight Perturbation Helps Robust Generalization , 2020, NeurIPS.
[22] David A. Wagner,et al. Obfuscated Gradients Give a False Sense of Security: Circumventing Defenses to Adversarial Examples , 2018, ICML.
[23] Timothy A. Mann,et al. Uncovering the Limits of Adversarial Training against Norm-Bounded Adversarial Examples , 2020, ArXiv.
[24] Bernhard Pfahringer,et al. Regularisation of neural networks by enforcing Lipschitz continuity , 2018, Machine Learning.
[25] James Bailey,et al. Characterizing Adversarial Subspaces Using Local Intrinsic Dimensionality , 2018, ICLR.
[26] Shu-Tao Xia,et al. Improving Adversarial Robustness via Channel-wise Activation Suppressing , 2021, ICLR.
[27] Yu-Gang Jiang,et al. Revisiting Adversarial Robustness Distillation: Robust Soft Labels Make Student Better , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).
[28] Gaurav Mittal,et al. An Empirical Study on the Robustness of NAS based Architectures , 2020, ArXiv.
[29] Aleksander Madry,et al. Adversarial Robustness as a Prior for Learned Representations , 2019 .
[30] Michael I. Jordan,et al. Theoretically Principled Trade-off between Robustness and Accuracy , 2019, ICML.
[31] Bin Dong,et al. You Only Propagate Once: Accelerating Adversarial Training via Maximal Principle , 2019, NeurIPS.
[32] Ohad Shamir,et al. Depth-Width Tradeoffs in Approximating Natural Functions with Neural Networks , 2016, ICML.
[33] James Bailey,et al. On the Convergence and Robustness of Adversarial Training , 2021, ICML.
[34] Aleksander Madry,et al. Adversarially Robust Generalization Requires More Data , 2018, NeurIPS.
[35] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[36] Quoc V. Le,et al. Adversarial Examples Improve Image Recognition , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[37] M. Rudelson,et al. Non-asymptotic theory of random matrices: extreme singular values , 2010, 1003.2990.
[38] Xiangning Chen,et al. Stabilizing Differentiable Architecture Search via Perturbation-based Regularization , 2020, ICML.
[39] Cyrus Rashtchian,et al. A Closer Look at Accuracy vs. Robustness , 2020, NeurIPS.
[40] Hang Su,et al. Boosting Adversarial Training with Hypersphere Embedding , 2020, NeurIPS.
[41] Zhouchen Lin,et al. Demystifying Adversarial Training via A Unified Probabilistic Framework , 2021 .
[42] J. Zico Kolter,et al. Overfitting in adversarially robust deep learning , 2020, ICML.
[43] Haifeng Qian,et al. L2-Nonexpansive Neural Networks , 2018, ICLR.
[44] Quanshi Zhang,et al. A Unified Approach to Interpreting and Boosting Adversarial Transferability , 2020, ICLR.
[45] Mohan S. Kankanhalli,et al. Attacks Which Do Not Kill Training Make Adversarial Learning Stronger , 2020, ICML.
[46] Colin Raffel,et al. Thermometer Encoding: One Hot Way To Resist Adversarial Examples , 2018, ICLR.
[47] Ludwig Schmidt,et al. Unlabeled Data Improves Adversarial Robustness , 2019, NeurIPS.
[48] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[49] Xingjun Ma,et al. Imbalanced Gradients: A New Cause of Overestimated Adversarial Robustness , 2020, ArXiv.
[50] Yu Wang,et al. Multi-shot NAS for Discovering Adversarially Robust Convolutional Neural Architectures at Targeted Capacities , 2020, ArXiv.
[51] Frank Hutter,et al. SGDR: Stochastic Gradient Descent with Warm Restarts , 2016, ICLR.
[52] J. Zico Kolter,et al. Fast is better than free: Revisiting adversarial training , 2020, ICLR.
[53] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[54] Matthias Hein,et al. Reliable evaluation of adversarial robustness with an ensemble of diverse parameter-free attacks , 2020, ICML.
[55] Pietro Perona,et al. Microsoft COCO: Common Objects in Context , 2014, ECCV.
[56] Jian Zhang,et al. SQuAD: 100,000+ Questions for Machine Comprehension of Text , 2016, EMNLP.
[57] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[58] Masashi Sugiyama,et al. Lipschitz-Margin Training: Scalable Certification of Perturbation Invariance for Deep Neural Networks , 2018, NeurIPS.
[59] Aleksander Madry,et al. Adversarial Examples Are Not Bugs, They Are Features , 2019, NeurIPS.
[60] Rui Xu,et al. When NAS Meets Robustness: In Search of Robust Architectures Against Adversarial Attacks , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[61] James Bailey,et al. Understanding Adversarial Attacks on Deep Learning Based Medical Image Analysis Systems , 2019, Pattern Recognit..
[62] Angela P. Schoellig,et al. An Analysis of the Expressiveness of Deep Neural Network Architectures Based on Their Lipschitz Constants , 2019, ArXiv.
[63] Quanshi Zhang,et al. Game-theoretic Understanding of Adversarially Learned Features , 2021, ArXiv.
[64] Nikos Komodakis,et al. Wide Residual Networks , 2016, BMVC.
[65] Aleksander Madry,et al. On Adaptive Attacks to Adversarial Example Defenses , 2020, NeurIPS.
[66] Jinfeng Yi,et al. Is Robustness the Cost of Accuracy? - A Comprehensive Study on the Robustness of 18 Deep Image Classification Models , 2018, ECCV.
[67] Alan L. Yuille,et al. Mitigating adversarial effects through randomization , 2017, ICLR.
[68] John Duchi,et al. Understanding and Mitigating the Tradeoff Between Robustness and Accuracy , 2020, ICML.
[69] Ruitong Huang,et al. Max-Margin Adversarial (MMA) Training: Direct Input Space Margin Maximization through Adversarial Training , 2018, ICLR.
[70] Larry S. Davis,et al. Adversarial Training for Free! , 2019, NeurIPS.
[71] Yan Wang,et al. RobustART: Benchmarking Robustness on Architecture Design and Training Techniques , 2021, ArXiv.
[72] Moustapha Cissé,et al. Parseval Networks: Improving Robustness to Adversarial Examples , 2017, ICML.
[73] Atul Prakash,et al. Robust Physical-World Attacks on Deep Learning Visual Classification , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[74] Ning Chen,et al. Rethinking Softmax Cross-Entropy Loss for Adversarial Robustness , 2019, ICLR.
[75] Moustapha Cissé,et al. Countering Adversarial Images using Input Transformations , 2018, ICLR.
[76] Li Fei-Fei,et al. ImageNet: A large-scale hierarchical image database , 2009, CVPR.
[77] Yiming Yang,et al. DARTS: Differentiable Architecture Search , 2018, ICLR.
[78] Simon Lucey,et al. Architectural Adversarial Robustness: The Case for Deep Pursuit , 2020, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[79] Ashish Kapoor,et al. Do Adversarially Robust ImageNet Models Transfer Better? , 2020, NeurIPS.
[80] James Bailey,et al. Adversarial Camouflage: Hiding Physical-World Attacks With Natural Styles , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[81] Yunhe Wang,et al. Adversarially Robust Neural Architectures , 2020, ArXiv.
[82] Hang Su,et al. Bag of Tricks for Adversarial Training , 2020, ICLR.
[83] Gang Niu,et al. Geometry-aware Instance-reweighted Adversarial Training , 2021, ICLR.
[84] Po-Sen Huang,et al. Are Labels Required for Improving Adversarial Robustness? , 2019, NeurIPS.
[85] Cho-Jui Hsieh,et al. Self-Progressing Robust Training , 2020, ArXiv.