暂无分享,去创建一个
Tom Goldstein | Christoph Studer | Ahmed Abdelkader | Chen Zhu | Renkun Ni | Ping-Yeh Chiang | T. Goldstein | Ping-Yeh Chiang | Renkun Ni | Chen Zhu | Christoph Studer | Ahmed Abdelkader | Ping-yeh Chiang
[1] Yizheng Chen,et al. Enhancing Gradient-based Attacks with Symbolic Intervals , 2019, ArXiv.
[2] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[3] J. Zico Kolter,et al. Certified Adversarial Robustness via Randomized Smoothing , 2019, ICML.
[4] Sergey Ioffe,et al. Rethinking the Inception Architecture for Computer Vision , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[5] J. Zico Kolter,et al. Provable defenses against adversarial examples via the convex outer adversarial polytope , 2017, ICML.
[6] J. Zico Kolter,et al. Wasserstein Adversarial Examples via Projected Sinkhorn Iterations , 2019, ICML.
[7] David A. Wagner,et al. Obfuscated Gradients Give a False Sense of Security: Circumventing Defenses to Adversarial Examples , 2018, ICML.
[8] Timothy A. Mann,et al. On the Effectiveness of Interval Bound Propagation for Training Verifiably Robust Models , 2018, ArXiv.
[9] Tom Goldstein,et al. Making an Invisibility Cloak: Real World Adversarial Attacks on Object Detectors , 2020, ECCV.
[10] Soheil Feizi,et al. Functional Adversarial Attacks , 2019, NeurIPS.
[11] Aleksander Madry,et al. A Rotation and a Translation Suffice: Fooling CNNs with Simple Transformations , 2017, ArXiv.
[12] Thomas Brox,et al. U-Net: Convolutional Networks for Biomedical Image Segmentation , 2015, MICCAI.
[13] Li Fei-Fei,et al. ImageNet: A large-scale hierarchical image database , 2009, CVPR.
[14] Seyed-Mohsen Moosavi-Dezfooli,et al. SparseFool: A Few Pixels Make a Big Difference , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[15] Cho-Jui Hsieh,et al. Towards Stable and Efficient Training of Verifiably Robust Neural Networks , 2019, ICLR.
[16] Yoav Goldberg,et al. LaVAN: Localized and Visible Adversarial Noise , 2018, ICML.
[17] Lujo Bauer,et al. Adversarial Generative Nets: Neural Network Attacks on State-of-the-Art Face Recognition , 2018, ArXiv.
[18] Matthew Mirman,et al. Differentiable Abstract Interpretation for Provably Robust Neural Networks , 2018, ICML.
[19] Aleksander Madry,et al. Exploring the Landscape of Spatial Robustness , 2017, ICML.
[20] Inderjit S. Dhillon,et al. Towards Fast Computation of Certified Robustness for ReLU Networks , 2018, ICML.
[21] Salman Khan,et al. Local Gradients Smoothing: Defense Against Localized Adversarial Attacks , 2018, 2019 IEEE Winter Conference on Applications of Computer Vision (WACV).
[22] Andrew Zisserman,et al. Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.
[23] Samy Bengio,et al. Adversarial Machine Learning at Scale , 2016, ICLR.
[24] Alexander Levine,et al. Robustness Certificates for Sparse Adversarial Attacks by Randomized Ablation , 2019, AAAI.
[25] Tom Goldstein,et al. Are adversarial examples inevitable? , 2018, ICLR.
[26] Jamie Hayes,et al. On Visible Adversarial Perturbations & Digital Watermarking , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).
[27] Mark Lee,et al. On Physical Adversarial Patches for Object Detection , 2019, ArXiv.
[28] Dan Boneh,et al. Adversarial Training and Robustness for Multiple Perturbations , 2019, NeurIPS.
[29] Larry S. Davis,et al. Adversarial Training for Free! , 2019, NeurIPS.
[30] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.