暂无分享,去创建一个
Zheng Zhang | Ting Wang | Shouling Ji | Ren Pang | Peng Cheng | Zhaohan Xi | Xiangshan Gao
[1] Vitaly Shmatikov,et al. Humpty Dumpty: Controlling Word Meanings via Corpus Poisoning , 2020, 2020 IEEE Symposium on Security and Privacy (SP).
[2] Graham Neubig,et al. Weight Poisoning Attacks on Pretrained Models , 2020, ACL.
[3] Kilian Q. Weinberger,et al. Densely Connected Convolutional Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[4] Benjamin Edwards,et al. Detecting Backdoor Attacks on Deep Neural Networks by Activation Clustering , 2018, SafeAI@AAAI.
[5] Mingjie Sun,et al. Poisoned classifiers are not only backdoored, they are fundamentally broken , 2020, ArXiv.
[6] Shanshan Peng,et al. Model Agnostic Defence Against Backdoor Attacks in Machine Learning , 2019, IEEE Transactions on Reliability.
[7] Aleksander Madry,et al. Clean-Label Backdoor Attacks , 2018 .
[8] Xiangyu Zhang,et al. ABS: Scanning Neural Networks for Back-doors by Artificial Brain Stimulation , 2019, CCS.
[9] Andrew Zisserman,et al. Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.
[10] Damith C. Ranasinghe,et al. Februus: Input Purification Defense Against Trojan Attacks on Deep Neural Network Systems , 2020, ACSAC.
[11] Brendan Dolan-Gavitt,et al. BadNets: Identifying Vulnerabilities in the Machine Learning Model Supply Chain , 2017, ArXiv.
[12] Michael Backes,et al. Dynamic Backdoor Attacks Against Machine Learning Models , 2020, ArXiv.
[13] D. Sculley,et al. Hidden Technical Debt in Machine Learning Systems , 2015, NIPS.
[14] Tudor Dumitras,et al. When Does Machine Learning FAIL? Generalized Transferability for Evasion and Poisoning Attacks , 2018, USENIX Security Symposium.
[15] Takeshi Fujino,et al. Disabling Backdoor and Identifying Poison Data by using Knowledge Distillation in Backdoor Attacks on Deep Neural Networks , 2020, AISec@CCS.
[16] Xiangyu Zhang,et al. Attacks Meet Interpretability: Attribute-steered Detection of Adversarial Samples , 2018, NeurIPS.
[17] Dawn Xiaodong Song,et al. Targeted Backdoor Attacks on Deep Learning Systems Using Data Poisoning , 2017, ArXiv.
[18] Kiran Karra,et al. The TrojAI Software Framework: An OpenSource tool for Embedding Trojans into Deep Learning Models , 2020, ArXiv.
[19] Pascal Frossard,et al. Analysis of universal adversarial perturbations , 2017, ArXiv.
[20] Yoshua Bengio,et al. How transferable are features in deep neural networks? , 2014, NIPS.
[21] Ben Y. Zhao,et al. Neural Cleanse: Identifying and Mitigating Backdoor Attacks in Neural Networks , 2019, 2019 IEEE Symposium on Security and Privacy (SP).
[22] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[23] Bo Li,et al. RAB: Provable Robustness Against Backdoor Attacks , 2020, 2023 IEEE Symposium on Security and Privacy (SP).
[24] Vitaly Shmatikov,et al. Blind Backdoors in Deep Learning Models , 2020, USENIX Security Symposium.
[25] J. Zico Kolter,et al. Certified Adversarial Robustness via Randomized Smoothing , 2019, ICML.
[26] Sebastian Thrun,et al. Dermatologist-level classification of skin cancer with deep neural networks , 2017, Nature.
[27] Yoshua Bengio,et al. Generative Adversarial Networks , 2014, ArXiv.
[28] Johannes Stallkamp,et al. Man vs. computer: Benchmarking machine learning algorithms for traffic sign recognition , 2012, Neural Networks.
[29] Jerry Li,et al. Spectral Signatures in Backdoor Attacks , 2018, NeurIPS.
[30] Li Fei-Fei,et al. ImageNet: A large-scale hierarchical image database , 2009, CVPR.
[31] Yong Jiang,et al. Backdoor Learning: A Survey , 2020, IEEE transactions on neural networks and learning systems.
[32] Andrea Vedaldi,et al. Interpretable Explanations of Black Boxes by Meaningful Perturbation , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[33] Jishen Zhao,et al. DeepInspect: A Black-box Trojan Detection and Mitigation Framework for Deep Neural Networks , 2019, IJCAI.
[34] Ankur Srivastava,et al. A Survey on Neural Trojans , 2020, 2020 21st International Symposium on Quality Electronic Design (ISQED).
[35] Fabio Roli,et al. Poisoning Adaptive Biometric Systems , 2012, SSPR/SPR.
[36] Minhui Xue,et al. Invisible Backdoor Attacks Against Deep Neural Networks , 2019, ArXiv.
[37] Damith Chinthana Ranasinghe,et al. STRIP: a defence against trojan attacks on deep neural networks , 2019, ACSAC.
[38] Ting Wang,et al. DEEPSEC: A Uniform Platform for Security Analysis of Deep Learning Model , 2019, 2019 IEEE Symposium on Security and Privacy (SP).
[39] Yanjun Qi,et al. Feature Squeezing: Detecting Adversarial Examples in Deep Neural Networks , 2017, NDSS.
[40] Yukun Yang,et al. Defending Neural Backdoors via Generative Distribution Modeling , 2019, NeurIPS.
[41] Nikita Borisov,et al. Detecting AI Trojans Using Meta Neural Analysis , 2019, 2021 IEEE Symposium on Security and Privacy (SP).
[42] Hamed Pirsiavash,et al. Hidden Trigger Backdoor Attacks , 2019, AAAI.
[43] Wen-Chuan Lee,et al. Trojaning Attack on Neural Networks , 2018, NDSS.
[44] Reza Shokri,et al. Bypassing Backdoor Detection Algorithms in Deep Learning , 2019, 2020 IEEE European Symposium on Security and Privacy (EuroS&P).
[45] Mani Srivastava,et al. NeuronInspect: Detecting Backdoors in Neural Networks via Output Explanations , 2019, ArXiv.
[46] Fan Yang,et al. An Embarrassingly Simple Approach for Trojan Attack in Deep Neural Networks , 2020, KDD.
[47] Tom Goldstein,et al. Transferable Clean-Label Poisoning Attacks on Deep Neural Nets , 2019, ICML.
[48] Percy Liang,et al. Certified Defenses for Data Poisoning Attacks , 2017, NIPS.
[49] Tudor Dumitras,et al. Poison Frogs! Targeted Clean-Label Poisoning Attacks on Neural Networks , 2018, NeurIPS.
[50] Abhishek Das,et al. Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization , 2016, 2017 IEEE International Conference on Computer Vision (ICCV).
[51] Ting Wang,et al. Model-Reuse Attacks on Deep Learning Systems , 2018, CCS.
[52] Yunfei Liu,et al. Reflection Backdoor: A Natural Backdoor Attack on Deep Neural Networks , 2020, ECCV.
[53] Xiapu Luo,et al. A Tale of Evil Twins: Adversarial Inputs versus Poisoned Models , 2019, CCS.
[54] Hao Chen,et al. MagNet: A Two-Pronged Defense against Adversarial Examples , 2017, CCS.
[55] Wenchao Li,et al. TrojDRL: Trojan Attacks on Deep Reinforcement Learning Agents , 2019, ArXiv.
[56] Omkar M. Parkhi,et al. VGGFace2: A Dataset for Recognising Faces across Pose and Age , 2017, 2018 13th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2018).
[57] Brendan Dolan-Gavitt,et al. Fine-Pruning: Defending Against Backdooring Attacks on Deep Neural Networks , 2018, RAID.
[58] James Bailey,et al. Skip Connections Matter: On the Transferability of Adversarial Examples Generated with ResNets , 2020, ICLR.
[59] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[60] Zheng Zhang,et al. Trojaning Language Models for Fun and Profit , 2020, 2021 IEEE European Symposium on Security and Privacy (EuroS&P).
[61] Ting Wang,et al. Interpretable Deep Learning under Fire , 2018, USENIX Security Symposium.
[62] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[63] Bo Li,et al. DBA: Distributed Backdoor Attacks against Federated Learning , 2020, ICLR.
[64] Vitaly Shmatikov,et al. How To Backdoor Federated Learning , 2018, AISTATS.
[65] Xiangyu Zhang,et al. Composite Backdoor Attack for Deep Neural Network by Mixing Existing Benign Features , 2020, CCS.
[66] Ben Y. Zhao,et al. Latent Backdoor Attacks on Deep Neural Networks , 2019, CCS.
[67] Wenbo Guo,et al. TABOR: A Highly Accurate Approach to Inspecting and Restoring Trojan Backdoors in AI Systems , 2019, ArXiv.
[68] Gang Wang,et al. LEMNA: Explaining Deep Learning based Security Applications , 2018, CCS.
[69] Anmin Fu,et al. Backdoor Attacks and Countermeasures on Deep Learning: A Comprehensive Review , 2020, ArXiv.