暂无分享,去创建一个
Wassim Hamidouche | Olivier Deforges | Ahmed Aldahdooh | W. Hamidouche | O. Déforges | Ahmed Aldahdooh
[1] Zhuowen Tu,et al. Aggregated Residual Transformations for Deep Neural Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[2] Ajmal Mian,et al. Threat of Adversarial Attacks on Deep Learning in Computer Vision: A Survey , 2018, IEEE Access.
[3] Wassim Hamidouche,et al. Adversarial example detection for DNN models: a review and experimental comparison , 2021, Artificial Intelligence Review.
[4] Chen Sun,et al. Revisiting Unreasonable Effectiveness of Data in Deep Learning Era , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[5] Yanjun Qi,et al. Feature Squeezing: Detecting Adversarial Examples in Deep Neural Networks , 2017, NDSS.
[6] Kaiming He,et al. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks , 2015, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[7] Samy Bengio,et al. Adversarial examples in the physical world , 2016, ICLR.
[8] Shiv Ram Dubey,et al. Color Channel Perturbation Attacks for Fooling Convolutional Neural Networks and A Defense Against Such Attacks , 2020, IEEE Transactions on Artificial Intelligence.
[9] Lucas Beyer,et al. Big Transfer (BiT): General Visual Representation Learning , 2020, ECCV.
[10] Matthias Hein,et al. Minimally distorted Adversarial Examples with a Fast Adaptive Boundary Attack , 2019, ICML.
[11] Pan He,et al. Adversarial Examples: Attacks and Defenses for Deep Learning , 2017, IEEE Transactions on Neural Networks and Learning Systems.
[12] Zoubin Ghahramani,et al. A study of the effect of JPG compression on adversarial images , 2016, ArXiv.
[13] Logan Engstrom,et al. Synthesizing Robust Adversarial Examples , 2017, ICML.
[14] Marcus A. Brubaker,et al. On the Effectiveness of Low Frequency Perturbations , 2019, IJCAI.
[15] David A. Wagner,et al. Obfuscated Gradients Give a False Sense of Security: Circumventing Defenses to Adversarial Examples , 2018, ICML.
[16] Jiliang Tang,et al. Adversarial Attacks and Defenses in Images, Graphs and Text: A Review , 2019, International Journal of Automation and Computing.
[17] Sergey Ioffe,et al. Rethinking the Inception Architecture for Computer Vision , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[18] Ananthram Swami,et al. The Limitations of Deep Learning in Adversarial Settings , 2015, 2016 IEEE European Symposium on Security and Privacy (EuroS&P).
[19] Jinfeng Yi,et al. On the Adversarial Robustness of Visual Transformers , 2021, ArXiv.
[20] David A. Wagner,et al. Towards Evaluating the Robustness of Neural Networks , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[21] Zewen Li,et al. A Survey of Convolutional Neural Networks: Analysis, Applications, and Prospects , 2020, IEEE Transactions on Neural Networks and Learning Systems.
[22] Seyed-Mohsen Moosavi-Dezfooli,et al. Hold me tight! Influence of discriminative features on deep network boundaries , 2020, NeurIPS.
[23] Abhishek Das,et al. Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization , 2016, 2017 IEEE International Conference on Computer Vision (ICCV).
[24] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[25] Seyed-Mohsen Moosavi-Dezfooli,et al. DeepFool: A Simple and Accurate Method to Fool Deep Neural Networks , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[26] Wassim Hamidouche,et al. Perceptual Evaluation of Adversarial Attacks for CNN-based Image Classification , 2019, 2019 Eleventh International Conference on Quality of Multimedia Experience (QoMEX).
[27] Nicolas Flammarion,et al. Square Attack: a query-efficient black-box adversarial attack via random search , 2020, ECCV.
[28] Aleksander Madry,et al. Adversarial Examples Are Not Bugs, They Are Features , 2019, NeurIPS.
[29] Eero P. Simoncelli,et al. Image quality assessment: from error visibility to structural similarity , 2004, IEEE Transactions on Image Processing.
[30] Ananthram Swami,et al. Distillation as a Defense to Adversarial Perturbations Against Deep Neural Networks , 2015, 2016 IEEE Symposium on Security and Privacy (SP).
[31] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[32] David Wagner,et al. Adversarial Examples Are Not Easily Detected: Bypassing Ten Detection Methods , 2017, AISec@CCS.
[33] N. Codella,et al. CvT: Introducing Convolutions to Vision Transformers , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).
[34] Debdeep Mukhopadhyay,et al. Adversarial Attacks and Defences: A Survey , 2018, ArXiv.
[35] Li Fei-Fei,et al. ImageNet: A large-scale hierarchical image database , 2009, CVPR.
[36] M. Nikolova. An Algorithm for Total Variation Minimization and Applications , 2004 .
[37] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[38] Marten van Dijk,et al. On the Robustness of Vision Transformers to Adversarial Examples , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).
[39] Enhua Wu,et al. Transformer in Transformer , 2021, NeurIPS.
[40] Bo Chen,et al. MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications , 2017, ArXiv.
[41] Eric C. Larson,et al. Most apparent distortion: full-reference image quality assessment and the role of strategy , 2010, J. Electronic Imaging.
[42] Sinan Kalkan,et al. Transformer-Encoder Detector Module: Using Context to Improve Robustness to Adversarial Attacks on Object Detection , 2020, 2020 25th International Conference on Pattern Recognition (ICPR).
[43] Jorge Nocedal,et al. On the limited memory BFGS method for large scale optimization , 1989, Math. Program..
[44] Moustapha Cissé,et al. Countering Adversarial Images using Input Transformations , 2018, ICLR.
[45] Li Chen,et al. Keeping the Bad Guys Out: Protecting and Vaccinating Deep Learning with JPEG Compression , 2017, ArXiv.
[46] Shuicheng Yan,et al. Tokens-to-Token ViT: Training Vision Transformers from Scratch on ImageNet , 2021, ArXiv.
[47] Andrew Zisserman,et al. Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.
[48] Jun Zhu,et al. Boosting Adversarial Attacks with Momentum , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[49] Georg Heigold,et al. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale , 2021, ICLR.
[50] Andreas Veit,et al. Understanding Robustness of Transformers for Image Classification , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).
[51] Pin-Yu Chen,et al. Vision Transformers are Robust Learners , 2021, AAAI.
[52] Seyed-Mohsen Moosavi-Dezfooli,et al. Universal Adversarial Perturbations , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[53] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[54] Kevin Gimpel,et al. Gaussian Error Linear Units (GELUs) , 2016 .
[55] Terrance E. Boult,et al. Assessing Threat of Adversarial Examples on Deep Neural Networks , 2016, 2016 15th IEEE International Conference on Machine Learning and Applications (ICMLA).
[56] I. Johnstone,et al. Ideal spatial adaptation by wavelet shrinkage , 1994 .
[57] Matthias Hein,et al. Reliable evaluation of adversarial robustness with an ensemble of diverse parameter-free attacks , 2020, ICML.
[58] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[59] Jinghui Chen,et al. RayS: A Ray Searching Method for Hard-label Adversarial Attack , 2020, KDD.