Neural Cleanse: Identifying and Mitigating Backdoor Attacks in Neural Networks
暂无分享,去创建一个
Ben Y. Zhao | Yuanshun Yao | Haitao Zheng | Bimal Viswanath | Huiying Li | Bolun Wang | Shawn Shan | Haitao Zheng | Yuanshun Yao | Shawn Shan | Huiying Li | Bolun Wang | Bimal Viswanath
[1] Wen-Chuan Lee,et al. Trojaning Attack on Neural Networks , 2018, NDSS.
[2] Percy Liang,et al. Certified Defenses for Data Poisoning Attacks , 2017, NIPS.
[3] Zhenkai Liang,et al. Neural Nets Can Learn Function Type Signatures From Binaries , 2017, USENIX Security Symposium.
[4] Dawn Xiaodong Song,et al. Adversarial Example Defenses: Ensembles of Weak Defenses are not Strong , 2017, ArXiv.
[5] Xiaogang Wang,et al. Deep Learning Face Representation from Predicting 10,000 Classes , 2014, 2014 IEEE Conference on Computer Vision and Pattern Recognition.
[6] Hao Chen,et al. MagNet: A Two-Pronged Defense against Adversarial Examples , 2017, CCS.
[7] Dawn Xiaodong Song,et al. Delving into Transferable Adversarial Examples and Black-box Attacks , 2016, ICLR.
[8] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[9] Ling Huang,et al. ANTIDOTE: understanding and defending against poisoning of anomaly detectors , 2009, IMC '09.
[10] David A. Wagner,et al. Towards Evaluating the Robustness of Neural Networks , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[11] Dawn Xiaodong Song,et al. Targeted Backdoor Attacks on Deep Learning Systems Using Data Poisoning , 2017, ArXiv.
[12] Junfeng Yang,et al. Efficient Repair of Polluted Machine Learning Systems via Causal Unlearning , 2018, AsiaCCS.
[13] Ananthram Swami,et al. Distillation as a Defense to Adversarial Perturbations Against Deep Neural Networks , 2015, 2016 IEEE Symposium on Security and Privacy (SP).
[14] Rui Peng,et al. Network Trimming: A Data-Driven Neuron Pruning Approach towards Efficient Deep Architectures , 2016, ArXiv.
[15] Alexander Binder,et al. On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation , 2015, PloS one.
[16] Chang Liu,et al. Manipulating Machine Learning: Poisoning Attacks and Countermeasures for Regression Learning , 2018, 2018 IEEE Symposium on Security and Privacy (SP).
[17] Johannes Stallkamp,et al. Man vs. computer: Benchmarking machine learning algorithms for traffic sign recognition , 2012, Neural Networks.
[18] Harini Kannan,et al. Adversarial Logit Pairing , 2018, NIPS 2018.
[19] Brendan Dolan-Gavitt,et al. Fine-Pruning: Defending Against Backdooring Attacks on Deep Neural Networks , 2018, RAID.
[20] David A. Wagner,et al. MagNet and "Efficient Defenses Against Adversarial Attacks" are Not Robust to Adversarial Examples , 2017, ArXiv.
[21] Ben Y. Zhao,et al. With Great Training Comes Great Vulnerability: Practical Attacks against Transfer Learning , 2018, USENIX Security Symposium.
[22] F. Hampel. The Influence Curve and Its Role in Robust Estimation , 1974 .
[23] Qi Wei,et al. Hu-Fu: Hardware and Software Collaborative Attack Framework Against Neural Networks , 2018, 2018 IEEE Computer Society Annual Symposium on VLSI (ISVLSI).
[24] Ananthram Swami,et al. The Limitations of Deep Learning in Adversarial Settings , 2015, 2016 IEEE European Symposium on Security and Privacy (EuroS&P).
[25] Wenbo Guo,et al. Adversary Resistant Deep Neural Networks with an Application to Malware Detection , 2016, KDD.
[26] Gang Wang,et al. LEMNA: Explaining Deep Learning based Security Applications , 2018, CCS.
[27] Dawn Xiaodong Song,et al. Recognizing Functions in Binaries with Neural Networks , 2015, USENIX Security Symposium.
[28] Yingjie Lao,et al. Hardware Trojan Attacks on Neural Networks , 2018, ArXiv.
[29] Konrad Rieck,et al. DREBIN: Effective and Explainable Detection of Android Malware in Your Pocket , 2014, NDSS.
[30] Yanjun Qi,et al. Feature Squeezing: Detecting Adversarial Examples in Deep Neural Networks , 2017, NDSS.
[31] Andrew Zisserman,et al. Deep Face Recognition , 2015, BMVC.
[32] Hervé Debar,et al. A neural network component for an intrusion detection system , 1992, Proceedings 1992 IEEE Computer Society Symposium on Research in Security and Privacy.
[33] P. Rousseeuw,et al. Alternatives to the Median Absolute Deviation , 1993 .
[34] David A. Wagner,et al. Defensive Distillation is Not Robust to Adversarial Examples , 2016, ArXiv.
[35] David A. Wagner,et al. Obfuscated Gradients Give a False Sense of Security: Circumventing Defenses to Adversarial Examples , 2018, ICML.
[36] Susmita Sur-Kolay,et al. Systematic Poisoning Attacks on and Defenses for Machine Learning in Healthcare , 2015, IEEE Journal of Biomedical and Health Informatics.
[37] Salvatore J. Stolfo,et al. Casting out Demons: Sanitizing Training Data for Anomaly Sensors , 2008, 2008 IEEE Symposium on Security and Privacy (sp 2008).
[38] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[39] Samy Bengio,et al. Adversarial Machine Learning at Scale , 2016, ICLR.
[40] Hod Lipson,et al. Understanding Neural Networks Through Deep Visualization , 2015, ArXiv.
[41] Brendan Dolan-Gavitt,et al. BadNets: Identifying Vulnerabilities in the Machine Learning Model Supply Chain , 2017, ArXiv.
[42] Scott Lundberg,et al. A Unified Approach to Interpreting Model Predictions , 2017, NIPS.
[43] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[44] Harris Drucker,et al. Learning algorithms for classification: A comparison on handwritten digit recognition , 1995 .
[45] Martín Abadi,et al. Adversarial Patch , 2017, ArXiv.
[46] Zachary Chase Lipton. The mythos of model interpretability , 2016, ACM Queue.
[47] Ankur Srivastava,et al. Neural Trojans , 2017, 2017 IEEE International Conference on Computer Design (ICCD).