暂无分享,去创建一个
Somesh Jha | Yingyu Liang | Xi Wu | Yang Guo | Jiefeng Chen
[1] Yann LeCun,et al. The mnist database of handwritten digits , 2005 .
[2] Suman Jana,et al. HYDRA: Pruning Adversarially Robust Neural Networks , 2020, NeurIPS.
[3] Kimin Lee,et al. Using Pre-Training Can Improve Model Robustness and Uncertainty , 2019, ICML.
[4] Seyed-Mohsen Moosavi-Dezfooli,et al. DeepFool: A Simple and Accurate Method to Fool Deep Neural Networks , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[5] François Laviolette,et al. Domain-Adversarial Neural Networks , 2014, ArXiv.
[6] Ludwig Schmidt,et al. Unlabeled Data Improves Adversarial Robustness , 2019, NeurIPS.
[7] J. Zico Kolter,et al. Certified Adversarial Robustness via Randomized Smoothing , 2019, ICML.
[8] Vladimir Vapnik,et al. Statistical learning theory , 1998 .
[9] Dawn Xiaodong Song,et al. Delving into Transferable Adversarial Examples and Black-box Attacks , 2016, ICLR.
[10] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[11] James Bailey,et al. Improving Adversarial Robustness Requires Revisiting Misclassified Examples , 2020, ICLR.
[12] 俊一 甘利. 5分で分かる!? 有名論文ナナメ読み:Jacot, Arthor, Gabriel, Franck and Hongler, Clement : Neural Tangent Kernel : Convergence and Generalization in Neural Networks , 2020 .
[13] Trevor Darrell,et al. Fighting Gradients with Gradients: Dynamic Defenses against Adversarial Attacks , 2021, ArXiv.
[14] Matthias Hein,et al. Reliable evaluation of adversarial robustness with an ensemble of diverse parameter-free attacks , 2020, ICML.
[15] John C. Duchi,et al. Certifying Some Distributional Robustness with Principled Adversarial Training , 2017, ICLR.
[16] Ching-Yao Chuang,et al. Estimating Generalization under Distribution Shifts via Domain-Invariant Representations , 2020, ICML.
[17] Michael I. Jordan,et al. Theoretically Principled Trade-off between Robustness and Accuracy , 2019, ICML.
[18] Johannes Stallkamp,et al. Man vs. computer: Benchmarking machine learning algorithms for traffic sign recognition , 2012, Neural Networks.
[19] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[20] Samy Bengio,et al. Adversarial examples in the physical world , 2016, ICLR.
[21] J. Zico Kolter,et al. Fast is better than free: Revisiting adversarial training , 2020, ICLR.
[22] Ian J. Goodfellow,et al. A Research Agenda: Dynamic Models to Defend Against Correlated Attacks , 2019, ArXiv.
[23] Aleksander Madry,et al. Adversarially Robust Generalization Requires More Data , 2018, NeurIPS.
[24] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[25] Xuancheng Ren,et al. An Adaptive and Momental Bound Method for Stochastic Learning , 2019, ArXiv.
[26] Shan-Hung Wu,et al. Adversarial Robustness via Runtime Masking and Cleansing , 2020, ICML.
[27] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[28] Yisen Wang,et al. Adversarial Weight Perturbation Helps Robust Generalization , 2020, NeurIPS.
[29] Yael Tauman Kalai,et al. Beyond Perturbations: Learning Guarantees with Arbitrary Adversarial Test Examples , 2020, NeurIPS.
[30] Ruitong Huang,et al. Max-Margin Adversarial (MMA) Training: Direct Input Space Margin Maximization through Adversarial Training , 2018, ICLR.
[31] Diane J. Cook,et al. A Survey of Unsupervised Deep Domain Adaptation , 2018, ACM Trans. Intell. Syst. Technol..
[32] David A. Wagner,et al. Obfuscated Gradients Give a False Sense of Security: Circumventing Defenses to Adversarial Examples , 2018, ICML.
[33] Patrice Marcotte,et al. An overview of bilevel optimization , 2007, Ann. Oper. Res..
[34] Aleksander Madry,et al. On Adaptive Attacks to Adversarial Example Defenses , 2020, NeurIPS.
[35] François Laviolette,et al. Domain-Adversarial Training of Neural Networks , 2015, J. Mach. Learn. Res..
[36] Thomas G. Dietterich,et al. Benchmarking Neural Network Robustness to Common Corruptions and Perturbations , 2018, ICLR.
[37] David A. Wagner,et al. Towards Evaluating the Robustness of Neural Networks , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[38] Dan Boneh,et al. The Space of Transferable Adversarial Examples , 2017, ArXiv.
[39] David Duvenaud,et al. Stochastic Hyperparameter Optimization through Hypernetworks , 2018, ArXiv.