暂无分享,去创建一个
Anima Anandkumar | De-An Huang | Homanga Bharadhwaj | Chaowei Xiao | Animesh Garg | Anima Anandkumar | Animesh Garg | De-An Huang | Homanga Bharadhwaj | Chaowei Xiao | Animesh Garg
[1] Li Fei-Fei,et al. ImageNet: A large-scale hierarchical image database , 2009, CVPR.
[2] Aleksander Madry,et al. On Evaluating Adversarial Robustness , 2019, ArXiv.
[3] Aditi Raghunathan,et al. Semidefinite relaxations for certifying robustness to adversarial examples , 2018, NeurIPS.
[4] Jinfeng Yi,et al. Attacking Visual Language Grounding with Adversarial Examples: A Case Study on Neural Image Captioning , 2017, ACL.
[5] Clifton A. Ericson,et al. Failure Mode and Effects Analysis , 2005 .
[6] Inioluwa Deborah Raji,et al. Model Cards for Model Reporting , 2018, FAT.
[7] Andrew Y. Ng,et al. CheXNet: Radiologist-Level Pneumonia Detection on Chest X-Rays with Deep Learning , 2017, ArXiv.
[8] Alexei A. Efros,et al. Colorful Image Colorization , 2016, ECCV.
[9] Mingyan Liu,et al. Characterizing Adversarial Examples Based on Spatial Consistency Information for Semantic Segmentation , 2018, ECCV.
[10] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[11] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[12] Pushmeet Kohli,et al. Efficient Neural Network Verification with Exactness Characterization , 2019, UAI.
[13] Rama Chellappa,et al. Defense-GAN: Protecting Classifiers Against Adversarial Attacks Using Generative Models , 2018, ICLR.
[14] James Bailey,et al. Characterizing Adversarial Subspaces Using Local Intrinsic Dimensionality , 2018, ICLR.
[15] Bolei Zhou,et al. Understanding the role of individual units in a deep neural network , 2020, Proceedings of the National Academy of Sciences.
[16] Jaakko Lehtinen,et al. Progressive Growing of GANs for Improved Quality, Stability, and Variation , 2017, ICLR.
[17] Deniz Erdogmus,et al. Structured Adversarial Attack: Towards General Implementation and Better Interpretability , 2018, ICLR.
[18] Cho-Jui Hsieh,et al. Automatic Perturbation Analysis for Scalable Certified Robustness and Beyond , 2020, NeurIPS.
[19] Inderjit S. Dhillon,et al. The Limitations of Adversarial Training and the Blind-Spot Attack , 2019, ICLR.
[20] Aditi Raghunathan,et al. Certified Robustness to Adversarial Word Substitutions , 2019, EMNLP.
[21] Atul Prakash,et al. Robust Physical-World Attacks on Deep Learning Visual Classification , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[22] Chun-Liang Li,et al. On Completeness-aware Concept-Based Explanations in Deep Neural Networks , 2020, NeurIPS.
[23] Yinda Zhang,et al. LSUN: Construction of a Large-scale Image Dataset using Deep Learning with Humans in the Loop , 2015, ArXiv.
[24] David A. Wagner,et al. Towards Evaluating the Robustness of Neural Networks , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[25] Sameer Singh,et al. Generating Natural Adversarial Examples , 2017, ICLR.
[26] Xiaogang Wang,et al. Deep Learning Face Attributes in the Wild , 2014, 2015 IEEE International Conference on Computer Vision (ICCV).
[27] Bolei Zhou,et al. Closed-Form Factorization of Latent Semantics in GANs , 2020, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[28] David A. Wagner,et al. Obfuscated Gradients Give a False Sense of Security: Circumventing Defenses to Adversarial Examples , 2018, ICML.
[29] Pin-Yu Chen,et al. Towards Verifying Robustness of Neural Networks Against A Family of Semantic Perturbations , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[30] Bolei Zhou,et al. Semantic photo manipulation with a generative image prior , 2019, ACM Trans. Graph..
[31] Timnit Gebru,et al. Datasheets for datasets , 2018, Commun. ACM.
[32] Fei-Fei Li,et al. ImageNet: A large-scale hierarchical image database , 2009, 2009 IEEE Conference on Computer Vision and Pattern Recognition.
[33] Alexei A. Efros,et al. Image-to-Image Translation with Conditional Adversarial Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[34] Jinfeng Yi,et al. AdvIT: Adversarial Frames Identifier Based on Temporal Consistency in Videos , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[35] Yifan Yu,et al. CheXpert: A Large Chest Radiograph Dataset with Uncertainty Labels and Expert Comparison , 2019, AAAI.
[36] Deli Zhao,et al. In-Domain GAN Inversion for Real Image Editing , 2020, ECCV.
[37] Jeff Donahue,et al. Large Scale Adversarial Representation Learning , 2019, NeurIPS.
[38] J. Zico Kolter,et al. Learning perturbation sets for robust machine learning , 2020, ICLR.
[39] Bolei Zhou,et al. InterFaceGAN: Interpreting the Disentangled Face Representation Learned by GANs , 2020, IEEE transactions on pattern analysis and machine intelligence.
[40] Matthew Mirman,et al. Fast and Effective Robustness Certification , 2018, NeurIPS.
[41] Quoc V. Le,et al. Smooth Adversarial Training , 2020, ArXiv.
[42] Bo Li,et al. MeshAdv: Adversarial Meshes for Visual Recognition , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[43] Cho-Jui Hsieh,et al. Towards Stable and Efficient Training of Verifiably Robust Neural Networks , 2019, ICLR.
[44] Inderjit S. Dhillon,et al. Towards Fast Computation of Certified Robustness for ReLU Networks , 2018, ICML.
[45] Demis Hassabis,et al. Improved protein structure prediction using potentials from deep learning , 2020, Nature.
[46] David Janz,et al. Learning to Drive in a Day , 2018, 2019 International Conference on Robotics and Automation (ICRA).
[47] Eric C Ford,et al. Evaluation of safety in a radiation oncology setting using failure mode and effects analysis. , 2009, International journal of radiation oncology, biology, physics.
[48] Cho-Jui Hsieh,et al. Efficient Neural Network Robustness Certification with General Activation Functions , 2018, NeurIPS.
[49] Timothy A. Mann,et al. On the Effectiveness of Interval Bound Propagation for Training Verifiably Robust Models , 2018, ArXiv.
[50] Renjie Liao,et al. SpAGNN: Spatially-Aware Graph Neural Networks for Relational Behavior Forecasting from Sensor Data , 2019, 2020 IEEE International Conference on Robotics and Automation (ICRA).
[51] Mislav Balunovic,et al. Certifying Geometric Robustness of Neural Networks , 2019, NeurIPS.
[52] Ananthram Swami,et al. Distillation as a Defense to Adversarial Perturbations Against Deep Neural Networks , 2015, 2016 IEEE Symposium on Security and Privacy (SP).
[53] Colin Raffel,et al. Thermometer Encoding: One Hot Way To Resist Adversarial Examples , 2018, ICLR.
[54] Junfeng Yang,et al. Efficient Formal Safety Analysis of Neural Networks , 2018, NeurIPS.
[55] Aditi Raghunathan,et al. Certified Defenses against Adversarial Examples , 2018, ICLR.
[56] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[57] Ilya Sutskever,et al. Learning Transferable Visual Models From Natural Language Supervision , 2021, ICML.
[58] Xiaoming Liu,et al. Disentangled Representation Learning GAN for Pose-Invariant Face Recognition , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[59] David de la Fuente,et al. A decision support system for applying failure mode and effects analysis , 2002 .
[60] Timo Aila,et al. A Style-Based Generator Architecture for Generative Adversarial Networks , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[61] Mingyan Liu,et al. Spatially Transformed Adversarial Examples , 2018, ICLR.
[62] P. Henriksen,et al. Efficient Neural Network Verification via Adaptive Refinement and Adversarial Search , 2020, ECAI.
[63] Jeff Donahue,et al. Large Scale GAN Training for High Fidelity Natural Image Synthesis , 2018, ICLR.
[64] Ronald M. Summers,et al. ChestX-ray: Hospital-Scale Chest X-ray Database and Benchmarks on Weakly Supervised Classification and Localization of Common Thorax Diseases , 2019, Deep Learning and Convolutional Neural Networks for Medical Imaging and Clinical Informatics.
[65] David Wagner,et al. Adversarial Examples Are Not Easily Detected: Bypassing Ten Detection Methods , 2017, AISec@CCS.
[66] Andre Esteva,et al. A guide to deep learning in healthcare , 2019, Nature Medicine.
[67] Mark Chen,et al. Language Models are Few-Shot Learners , 2020, NeurIPS.
[68] Forrest N. Iandola,et al. DenseNet: Implementing Efficient ConvNet Descriptor Pyramids , 2014, ArXiv.
[69] Fei Wang,et al. Deep learning for healthcare: review, opportunities and challenges , 2018, Briefings Bioinform..
[70] Xiang Li,et al. Adversarial Metric Learning , 2018, IJCAI.
[71] Mingyan Liu,et al. Generating Adversarial Examples with Adversarial Networks , 2018, IJCAI.
[72] Moustapha Cissé,et al. Countering Adversarial Images using Input Transformations , 2018, ICLR.
[73] Mykel J. Kochenderfer,et al. Algorithms for Verifying Deep Neural Networks , 2019, Found. Trends Optim..
[74] David Pfau,et al. Towards a Definition of Disentangled Representations , 2018, ArXiv.
[75] J. Zico Kolter,et al. Certified Adversarial Robustness via Randomized Smoothing , 2019, ICML.
[76] Yang Song,et al. PixelDefend: Leveraging Generative Models to Understand and Defend against Adversarial Examples , 2017, ICLR.
[77] Maximilian Baader,et al. Efficient Certification of Spatial Robustness , 2020, AAAI.