Connecting the Digital and Physical World: Improving the Robustness of Adversarial Attacks
暂无分享,去创建一个
Gang Wang | Yen-Chen Lin | Jia-Bin Huang | Steve T. K. Jan | Joseph C.E. Messou | G. Wang | Jia-Bin Huang | Yen-Chen Lin
[1] John B. Shoven,et al. I , Edinburgh Medical and Surgical Journal.
[2] Jan Kautz,et al. High-Resolution Image Synthesis and Semantic Manipulation with Conditional GANs , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[3] Dawn Xiaodong Song,et al. Delving into adversarial attacks on deep policies , 2017, ICLR.
[4] Jiajun Lu,et al. Adversarial Examples that Fool Detectors , 2017, ArXiv.
[5] Seyed-Mohsen Moosavi-Dezfooli,et al. DeepFool: A Simple and Accurate Method to Fool Deep Neural Networks , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[6] Antonio Criminisi,et al. Measuring Neural Net Robustness with Constraints , 2016, NIPS.
[7] Samy Bengio,et al. Adversarial examples in the physical world , 2016, ICLR.
[8] Vladlen Koltun,et al. Photographic Image Synthesis with Cascaded Refinement Networks , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[9] Tsuyoshi Murata,et al. {m , 1934, ACML.
[10] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[11] Prateek Mittal,et al. DARTS: Deceiving Autonomous Cars with Toxic Signs , 2018, ArXiv.
[12] Danna Zhou,et al. d. , 1934, Microbial pathogenesis.
[13] Facebook,et al. Houdini : Fooling Deep Structured Visual and Speech Recognition Models with Adversarial Examples , 2017 .
[14] Ming-Yu Liu,et al. Tactics of Adversarial Attack on Deep Reinforcement Learning Agents , 2017, IJCAI.
[15] Bo Chen,et al. MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications , 2017, ArXiv.
[16] Lujo Bauer,et al. Accessorize to a Crime: Real and Stealthy Attacks on State-of-the-Art Face Recognition , 2016, CCS.
[17] Logan Engstrom,et al. Synthesizing Robust Adversarial Examples , 2017, ICML.
[18] Alexei A. Efros,et al. The Unreasonable Effectiveness of Deep Features as a Perceptual Metric , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[19] Thomas Brox,et al. Adversarial Examples for Semantic Image Segmentation , 2017, ICLR.
[20] Ananthram Swami,et al. Practical Black-Box Attacks against Machine Learning , 2016, AsiaCCS.
[21] Sandy H. Huang,et al. Adversarial Attacks on Neural Network Policies , 2017, ICLR.
[22] Ananthram Swami,et al. The Limitations of Deep Learning in Adversarial Settings , 2015, 2016 IEEE European Symposium on Security and Privacy (EuroS&P).
[23] Michael S. Bernstein,et al. ImageNet Large Scale Visual Recognition Challenge , 2014, International Journal of Computer Vision.
[24] François Chollet,et al. Xception: Deep Learning with Depthwise Separable Convolutions , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[25] Terrance E. Boult,et al. Adversarial Diversity and Hard Positive Generation , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).
[26] Alan L. Yuille,et al. Adversarial Examples for Semantic Segmentation and Object Detection , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[27] Martín Abadi,et al. Adversarial Patch , 2017, ArXiv.
[28] Eero P. Simoncelli,et al. Image quality assessment: from error visibility to structural similarity , 2004, IEEE Transactions on Image Processing.
[29] Alexei A. Efros,et al. Image-to-Image Translation with Conditional Adversarial Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[30] Yanjun Qi,et al. Feature Squeezing: Detecting Adversarial Examples in Deep Neural Networks , 2017, NDSS.
[31] David A. Forsyth,et al. NO Need to Worry about Adversarial Examples in Object Detection in Autonomous Vehicles , 2017, ArXiv.
[32] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[33] Atul Prakash,et al. Note on Attacking Object Detectors with Adversarial Stickers , 2017, ArXiv.
[34] 拓海 杉山,et al. “Unpaired Image-to-Image Translation using Cycle-Consistent Adversarial Networks”の学習報告 , 2017 .
[35] Atul Prakash,et al. Robust Physical-World Attacks on Machine Learning Models , 2017, ArXiv.
[36] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[37] Sergey Ioffe,et al. Rethinking the Inception Architecture for Computer Vision , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).