Adversarial Robustness Curves
暂无分享,去创建一个
[1] David A. Wagner,et al. Towards Evaluating the Robustness of Neural Networks , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[2] Luca Rigazio,et al. Towards Deep Neural Network Architectures Robust to Adversarial Examples , 2014, ICLR.
[3] Ananthram Swami,et al. Distillation as a Defense to Adversarial Perturbations Against Deep Neural Networks , 2015, 2016 IEEE Symposium on Security and Privacy (SP).
[4] Ananthram Swami,et al. Practical Black-Box Attacks against Machine Learning , 2016, AsiaCCS.
[5] Aleksander Madry,et al. Robustness May Be at Odds with Accuracy , 2018, ICLR.
[6] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[7] Dale Schuurmans,et al. Learning with a Strong Adversary , 2015, ArXiv.
[8] Preetum Nakkiran,et al. Adversarial Robustness May Be at Odds With Simplicity , 2019, ArXiv.
[9] Samy Bengio,et al. Adversarial examples in the physical world , 2016, ICLR.
[10] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[11] Heiko Wersing,et al. Adversarial attacks hidden in plain sight , 2019, IDA.
[12] Ananthram Swami,et al. Practical Black-Box Attacks against Deep Learning Systems using Adversarial Examples , 2016, ArXiv.
[13] Kouichi Sakurai,et al. One Pixel Attack for Fooling Deep Neural Networks , 2017, IEEE Transactions on Evolutionary Computation.
[14] Antonio Criminisi,et al. Measuring Neural Net Robustness with Constraints , 2016, NIPS.
[15] Samy Bengio,et al. Adversarial Machine Learning at Scale , 2016, ICLR.