暂无分享,去创建一个
[1] Nikos Komodakis,et al. Wide Residual Networks , 2016, BMVC.
[2] Matthias Hein,et al. Reliable evaluation of adversarial robustness with an ensemble of diverse parameter-free attacks , 2020, ICML.
[3] Michael I. Jordan,et al. Theoretically Principled Trade-off between Robustness and Accuracy , 2019, ICML.
[4] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[5] Bin Dong,et al. You Only Propagate Once: Accelerating Adversarial Training via Maximal Principle , 2019, NeurIPS.
[6] Aleksander Madry,et al. Computer Vision with a Single (Robust) Classifier , 2019, NeurIPS 2019.
[7] Yvo Desmedt,et al. Man-in-the-Middle Attack , 2005, Encyclopedia of Cryptography and Security.
[8] J. Zico Kolter,et al. Provable defenses against adversarial examples via the convex outer adversarial polytope , 2017, ICML.
[9] Ilya P. Razenshteyn,et al. Randomized Smoothing of All Shapes and Sizes , 2020, ICML.
[10] Michael S. Bernstein,et al. ImageNet Large Scale Visual Recognition Challenge , 2014, International Journal of Computer Vision.
[11] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[12] Suman Jana,et al. Certified Robustness to Adversarial Examples with Differential Privacy , 2018, 2019 IEEE Symposium on Security and Privacy (SP).
[13] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[14] Greg Yang,et al. Provably Robust Deep Learning via Adversarially Trained Smoothed Classifiers , 2019, NeurIPS.
[15] Ben Y. Zhao,et al. Fawkes: Protecting Privacy against Unauthorized Deep Learning Models , 2020, USENIX Security Symposium.
[16] Sergey Levine,et al. Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks , 2017, ICML.
[17] Natalia Gimelshein,et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library , 2019, NeurIPS.
[18] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[19] Aditi Raghunathan,et al. Certified Defenses against Adversarial Examples , 2018, ICLR.
[20] Haichao Zhang,et al. Defense Against Adversarial Attacks Using Feature Scattering-based Adversarial Training , 2019, NeurIPS.
[21] Mohan S. Kankanhalli,et al. Attacks Which Do Not Kill Training Make Adversarial Learning Stronger , 2020, ICML.
[22] J. Zico Kolter,et al. Fast is better than free: Revisiting adversarial training , 2020, ICLR.
[23] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[24] Amit K. Roy-Chowdhury,et al. Adversarial Perturbations Against Real-Time Video Classification Systems , 2018, NDSS.
[25] Surya Nepal,et al. Man-in-the-Middle Attacks Against Machine Learning Classifiers Via Malicious Generative Models , 2019, IEEE Transactions on Dependable and Secure Computing.
[26] J. Zico Kolter,et al. Overfitting in adversarially robust deep learning , 2020, ICML.
[27] Seong Joon Oh,et al. Adversarial Image Perturbation for Privacy Protection A Game Theory Perspective , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[28] J. Zico Kolter,et al. Certified Adversarial Robustness via Randomized Smoothing , 2019, ICML.
[29] Ashish Kapoor,et al. Unadversarial Examples: Designing Objects for Robust Vision , 2020, NeurIPS.
[30] Micah Goldblum,et al. LowKey: Leveraging Adversarial Attacks to Protect Social Media Users from Facial Recognition , 2021, ICLR.
[31] Aleksander Madry,et al. Robustness May Be at Odds with Accuracy , 2018, ICLR.