暂无分享,去创建一个
R. Venkatesh Babu | Sravanti Addepalli | Arya Baburaj | Gaurang Sriramanan | Sravanti Addepalli | Gaurang Sriramanan | Arya Baburaj
[1] Nikos Komodakis,et al. Wide Residual Networks , 2016, BMVC.
[2] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[3] Li Fei-Fei,et al. ImageNet: A large-scale hierarchical image database , 2009, CVPR.
[4] Kimin Lee,et al. Using Pre-Training Can Improve Model Robustness and Uncertainty , 2019, ICML.
[5] Luiz Eduardo Soares de Oliveira,et al. Decoupling Direction and Norm for Efficient Gradient-Based L2 Adversarial Attacks and Defenses , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[6] Nicolas Flammarion,et al. Square Attack: a query-efficient black-box adversarial attack via random search , 2020, ECCV.
[7] R. Venkatesh Babu,et al. Regularizer to Mitigate Gradient Masking Effect During Single-Step Adversarial Training , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).
[8] Yann LeCun,et al. The mnist database of handwritten digits , 2005 .
[9] Pushmeet Kohli,et al. Adversarial Robustness through Local Linearization , 2019, NeurIPS.
[10] Ananthram Swami,et al. Practical Black-Box Attacks against Machine Learning , 2016, AsiaCCS.
[11] Alan L. Yuille,et al. Mitigating adversarial effects through randomization , 2017, ICLR.
[12] Yang Song,et al. PixelDefend: Leveraging Generative Models to Understand and Defend against Adversarial Examples , 2017, ICLR.
[13] Po-Sen Huang,et al. An Alternative Surrogate Loss for PGD-based Adversarial Testing , 2019, ArXiv.
[14] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[15] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[16] Matthias Hein,et al. Minimally distorted Adversarial Examples with a Fast Adaptive Boundary Attack , 2019, ICML.
[17] J. Zico Kolter,et al. Certified Adversarial Robustness via Randomized Smoothing , 2019, ICML.
[18] Samy Bengio,et al. Adversarial examples in the physical world , 2016, ICLR.
[19] Shai Shalev-Shwartz,et al. On Graduated Optimization for Stochastic Non-Convex Problems , 2015, ICML.
[20] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[21] R. Venkatesh Babu,et al. Towards Achieving Adversarial Robustness by Enforcing Feature Consistency Across Bit Planes , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[22] Matthias Hein,et al. Reliable evaluation of adversarial robustness with an ensemble of diverse parameter-free attacks , 2020, ICML.
[23] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[24] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[25] J. Zico Kolter,et al. Overfitting in adversarially robust deep learning , 2020, ICML.
[26] Samy Bengio,et al. Adversarial Machine Learning at Scale , 2016, ICLR.
[27] Aleksander Madry,et al. On Evaluating Adversarial Robustness , 2019, ArXiv.
[28] Philip Wolfe,et al. An algorithm for quadratic programming , 1956 .
[29] Colin Raffel,et al. Thermometer Encoding: One Hot Way To Resist Adversarial Examples , 2018, ICLR.
[30] Dan Boneh,et al. Ensemble Adversarial Training: Attacks and Defenses , 2017, ICLR.
[31] Ludwig Schmidt,et al. Unlabeled Data Improves Adversarial Robustness , 2019, NeurIPS.
[32] Pushmeet Kohli,et al. Adversarial Risk and the Dangers of Evaluating Against Weak Attacks , 2018, ICML.
[33] Moustapha Cissé,et al. Countering Adversarial Images using Input Transformations , 2018, ICLR.
[34] J. Zico Kolter,et al. Fast is better than free: Revisiting adversarial training , 2020, ICLR.
[35] Seyed-Mohsen Moosavi-Dezfooli,et al. Robustness via Curvature Regularization, and Vice Versa , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[36] David A. Wagner,et al. Towards Evaluating the Robustness of Neural Networks , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[37] Suman Jana,et al. HYDRA: Pruning Adversarially Robust Neural Networks , 2020, NeurIPS.
[38] David A. Wagner,et al. Obfuscated Gradients Give a False Sense of Security: Circumventing Defenses to Adversarial Examples , 2018, ICML.
[39] Larry S. Davis,et al. Adversarial Training for Free! , 2019, NeurIPS.
[40] Aleksander Madry,et al. Robustness May Be at Odds with Accuracy , 2018, ICLR.
[41] James Bailey,et al. Improving Adversarial Robustness Requires Revisiting Misclassified Examples , 2020, ICLR.