暂无分享,去创建一个
Alberto L. Sangiovanni-Vincentelli | Sanjit A. Seshia | Shromona Ghosh | Tommaso Dreossi | A. Sangiovanni-Vincentelli | S. Seshia | T. Dreossi | Shromona Ghosh
[1] Ashish Tiwari,et al. Output Range Analysis for Deep Feedforward Neural Networks , 2018, NFM.
[2] Xiaowei Huang,et al. Reachability Analysis of Deep Neural Networks with Provable Guarantees , 2018, IJCAI.
[3] Patrick D. McDaniel,et al. Transferability in Machine Learning: from Phenomena to Black-Box Attacks using Adversarial Samples , 2016, ArXiv.
[4] Matthew Wicker,et al. Feature-Guided Black-Box Safety Testing of Deep Neural Networks , 2017, TACAS.
[5] Blaine Nelson,et al. Poisoning Attacks against Support Vector Machines , 2012, ICML.
[6] Patrick D. McDaniel,et al. Making machine learning robust against adversarial inputs , 2018, Commun. ACM.
[7] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[8] J. Zico Kolter,et al. Provable defenses against adversarial examples via the convex outer adversarial polytope , 2017, ICML.
[9] Alberto L. Sangiovanni-Vincentelli,et al. Counterexample-Guided Data Augmentation , 2018, IJCAI.
[10] Ananthram Swami,et al. The Limitations of Deep Learning in Adversarial Settings , 2015, 2016 IEEE European Symposium on Security and Privacy (EuroS&P).
[11] Anna Philippou,et al. Tools and Algorithms for the Construction and Analysis of Systems , 2018, Lecture Notes in Computer Science.
[12] Ananthram Swami,et al. Practical Black-Box Attacks against Machine Learning , 2016, AsiaCCS.
[13] Logan Engstrom,et al. Synthesizing Robust Adversarial Examples , 2017, ICML.
[14] Junfeng Yang,et al. DeepXplore: Automated Whitebox Testing of Deep Learning Systems , 2017, SOSP.
[15] Valentina Zantedeschi,et al. Efficient Defenses Against Adversarial Attacks , 2017, AISec@CCS.
[16] Yanjun Qi,et al. Automatically Evading Classifiers: A Case Study on PDF Malware Classifiers , 2016, NDSS.
[17] Mykel J. Kochenderfer,et al. Reluplex: An Efficient SMT Solver for Verifying Deep Neural Networks , 2017, CAV.
[18] Hung Dang,et al. Evading Classifiers by Morphing in the Dark , 2017, CCS.
[19] Seyed-Mohsen Moosavi-Dezfooli,et al. DeepFool: A Simple and Accurate Method to Fool Deep Neural Networks , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[20] Sanjit A. Seshia,et al. Compositional Falsification of Cyber-Physical Systems with Machine Learning Components , 2017, NFM.
[21] Blaine Nelson,et al. Can machine learning be secure? , 2006, ASIACCS '06.
[22] Fan Zhang,et al. Stealing Machine Learning Models via Prediction APIs , 2016, USENIX Security Symposium.
[23] Michael P. Wellman,et al. Towards the Science of Security and Privacy in Machine Learning , 2016, ArXiv.
[24] Ananthram Swami,et al. Distillation as a Defense to Adversarial Perturbations Against Deep Neural Networks , 2015, 2016 IEEE Symposium on Security and Privacy (SP).
[25] David A. Wagner,et al. Towards Evaluating the Robustness of Neural Networks , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[26] Somesh Jha,et al. Semantic Adversarial Deep Learning , 2018, IEEE Design & Test.
[27] Jinfeng Yi,et al. ZOO: Zeroth Order Optimization Based Black-box Attacks to Deep Neural Networks without Training Substitute Models , 2017, AISec@CCS.
[28] Jinfeng Yi,et al. Evaluating the Robustness of Neural Networks: An Extreme Value Theory Approach , 2018, ICLR.
[29] Yanjun Qi,et al. Feature Squeezing: Detecting Adversarial Examples in Deep Neural Networks , 2017, NDSS.
[30] David A. Wagner,et al. Obfuscated Gradients Give a False Sense of Security: Circumventing Defenses to Adversarial Examples , 2018, ICML.
[31] Min Wu,et al. Safety Verification of Deep Neural Networks , 2016, CAV.
[32] Dawn Xiaodong Song,et al. Delving into Transferable Adversarial Examples and Black-box Attacks , 2016, ICLR.
[33] Pushmeet Kohli,et al. A Dual Approach to Scalable Verification of Deep Networks , 2018, UAI.
[34] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[35] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[36] Alberto L. Sangiovanni-Vincentelli,et al. Systematic Testing of Convolutional Neural Networks for Autonomous Driving , 2017, ArXiv.
[37] Sanjit A. Seshia,et al. Towards Verified Artificial Intelligence , 2016, ArXiv.