Boosting Adversarial Robustness From The Perspective of Effective Margin Regularization
暂无分享,去创建一个
[1] Sven Gowal,et al. Data Augmentation Can Improve Robustness , 2021, NeurIPS.
[2] Sven Gowal,et al. Improving Robustness using Generated Data , 2021, NeurIPS.
[3] Quanquan Gu,et al. Exploring Architectural Ingredients of Adversarially Robust Deep Neural Networks , 2021, NeurIPS.
[4] Martin Renqiang Min,et al. Towards Robustness of Deep Neural Networks via Regularization , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).
[5] Masashi Sugiyama,et al. Probabilistic Margins for Instance Reweighting in Adversarial Training , 2021, NeurIPS.
[6] Thomas Hofmann,et al. Uniform Convergence, Adversarial Spheres and a Simple Remedy , 2021, ICML.
[7] Jiaya Jia,et al. Learnable Boundary Guided Adversarial Training , 2020, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).
[8] Nicolas Flammarion,et al. RobustBench: a standardized adversarial robustness benchmark , 2020, NeurIPS Datasets and Benchmarks.
[9] Antoni B. Chan,et al. Improve Generalization and Robustness of Neural Networks via Weight Scale Shifting Invariant Regularizations , 2020, ArXiv.
[10] Quoc V. Le,et al. Smooth Adversarial Training , 2020, ArXiv.
[11] James Bailey,et al. Improving Adversarial Robustness Requires Revisiting Misclassified Examples , 2020, ICLR.
[12] Matthias Hein,et al. Reliable evaluation of adversarial robustness with an ensemble of diverse parameter-free attacks , 2020, ICML.
[13] J. Z. Kolter,et al. Overfitting in adversarially robust deep learning , 2020, ICML.
[14] Hang Su,et al. Boosting Adversarial Training with Hypersphere Embedding , 2020, NeurIPS.
[15] Francis Bach,et al. Implicit Bias of Gradient Descent for Wide Two-layer Neural Networks Trained with the Logistic Loss , 2020, COLT.
[16] Nicolas Flammarion,et al. Square Attack: a query-efficient black-box adversarial attack via random search , 2019, ECCV.
[17] Pushmeet Kohli,et al. Adversarial Robustness through Local Linearization , 2019, NeurIPS.
[18] Matthias Hein,et al. Minimally distorted Adversarial Examples with a Fast Adaptive Boundary Attack , 2019, ICML.
[19] Kaifeng Lyu,et al. Gradient Descent Maximizes the Margin of Homogeneous Neural Networks , 2019, ICLR.
[20] Ning Chen,et al. Rethinking Softmax Cross-Entropy Loss for Adversarial Robustness , 2019, ICLR.
[21] Aleksander Madry,et al. Adversarial Examples Are Not Bugs, They Are Features , 2019, NeurIPS.
[22] Thomas Hofmann,et al. The Odds are Odd: A Statistical Test for Detecting Adversarial Examples , 2019, ICML.
[23] Michael I. Jordan,et al. Theoretically Principled Trade-off between Robustness and Accuracy , 2019, ICML.
[24] Ruitong Huang,et al. Max-Margin Adversarial (MMA) Training: Direct Input Space Margin Maximization through Adversarial Training , 2018, ICLR.
[25] Wei Hu,et al. Algorithmic Regularization in Learning Deep Homogeneous Models: Layers are Automatically Balanced , 2018, NeurIPS.
[26] Aleksander Madry,et al. Robustness May Be at Odds with Accuracy , 2018, ICLR.
[27] Rama Chellappa,et al. Defense-GAN: Protecting Classifiers Against Adversarial Attacks Using Generative Models , 2018, ICLR.
[28] Kamyar Azizzadenesheli,et al. Stochastic Activation Pruning for Robust Adversarial Defense , 2018, ICLR.
[29] Colin Raffel,et al. Thermometer Encoding: One Hot Way To Resist Adversarial Examples , 2018, ICLR.
[30] David A. Wagner,et al. Obfuscated Gradients Give a False Sense of Security: Circumventing Defenses to Adversarial Examples , 2018, ICML.
[31] Jian Cheng,et al. Additive Margin Softmax for Face Verification , 2018, IEEE Signal Processing Letters.
[32] James Bailey,et al. Characterizing Adversarial Subspaces Using Local Intrinsic Dimensionality , 2018, ICLR.
[33] Andrew Slavin Ross,et al. Improving the Adversarial Robustness and Interpretability of Deep Neural Networks by Regularizing their Input Gradients , 2017, AAAI.
[34] Alan L. Yuille,et al. Mitigating adversarial effects through randomization , 2017, ICLR.
[35] Yang Song,et al. PixelDefend: Leveraging Generative Models to Understand and Defend against Adversarial Examples , 2017, ICLR.
[36] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[37] Jun Zhu,et al. Towards Robust Detection of Adversarial Examples , 2017, NeurIPS.
[38] Bhiksha Raj,et al. SphereFace: Deep Hypersphere Embedding for Face Recognition , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[39] Yanjun Qi,et al. Feature Squeezing: Detecting Adversarial Examples in Deep Neural Networks , 2017, NDSS.
[40] David A. Forsyth,et al. SafetyNet: Detecting and Rejecting Adversarial Examples Robustly , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[41] Samy Bengio,et al. Adversarial Machine Learning at Scale , 2016, ICLR.
[42] David A. Wagner,et al. Towards Evaluating the Robustness of Neural Networks , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[43] Meng Yang,et al. Large-Margin Softmax Loss for Convolutional Neural Networks , 2016, ICML.
[44] Nikos Komodakis,et al. Wide Residual Networks , 2016, BMVC.
[45] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[46] Seyed-Mohsen Moosavi-Dezfooli,et al. DeepFool: A Simple and Accurate Method to Fool Deep Neural Networks , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[47] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[48] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[49] Koby Crammer,et al. On the Algorithmic Implementation of Multiclass Kernel-based Vector Machines , 2002, J. Mach. Learn. Res..
[50] Anders Krogh,et al. A Simple Weight Decay Can Improve Generalization , 1991, NIPS.
[51] Moustapha Cissé,et al. Countering Adversarial Images using Input Transformations , 2018, ICLR.
[52] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[53] Yann LeCun,et al. The mnist database of handwritten digits , 2005 .