-
爱吃猫的鱼0于 2022年7月18日 02:45
[1] Andrew Y. Ng,et al. Reading Digits in Natural Images with Unsupervised Feature Learning , 2011 .
[2] Aleksander Madry,et al. Label-Consistent Backdoor Attacks , 2019, ArXiv.
[3] Wenbo Guo,et al. TABOR: A Highly Accurate Approach to Inspecting and Restoring Trojan Backdoors in AI Systems , 2019, ArXiv.
[4] J. H. Metzen,et al. Efficient Certified Defenses Against Patch Attacks on Image Classifiers , 2021, ICLR.
[5] Wen-Chuan Lee,et al. Trojaning Attack on Neural Networks , 2018, NDSS.
[6] Zheng Zhang,et al. Trojaning Language Models for Fun and Profit , 2020, 2021 IEEE European Symposium on Security and Privacy (EuroS&P).
[7] Chong Xiang,et al. PatchGuard: Provable Defense against Adversarial Patches Using Masks on Small Receptive Fields , 2020, ArXiv.
[8] Vitaly Shmatikov,et al. How To Backdoor Federated Learning , 2018, AISTATS.
[9] Bao Gia Doan,et al. Februus: Input Purification Defence Against Trojan Attacks on Deep Neural Network Systems , 2019, 1908.03369.
[10] Nan Duan,et al. Unicoder-VL: A Universal Encoder for Vision and Language by Cross-modal Pre-training , 2019, AAAI.
[11] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[12] Xiaoyu Cao,et al. Certified Robustness of Nearest Neighbors against Data Poisoning Attacks , 2020, ArXiv.
[13] Paolo Favaro,et al. Unsupervised Learning of Visual Representations by Solving Jigsaw Puzzles , 2016, ECCV.
[14] H. Pirsiavash,et al. Hidden Trigger Backdoor Attacks , 2019, AAAI.
[15] Geoffrey E. Hinton,et al. A Simple Framework for Contrastive Learning of Visual Representations , 2020, ICML.
[16] David A. Shamma,et al. YFCC100M , 2015, Commun. ACM.
[17] Pierre H. Richemond,et al. Bootstrap Your Own Latent: A New Approach to Self-Supervised Learning , 2020, NeurIPS.
[18] Benjamin Zi Hao Zhao,et al. Invisible Backdoor Attacks Against Deep Neural Networks , 2019, ArXiv.
[19] Ben Y. Zhao,et al. Neural Cleanse: Identifying and Mitigating Backdoor Attacks in Neural Networks , 2019, 2019 IEEE Symposium on Security and Privacy (SP).
[20] Yann LeCun,et al. Dimensionality Reduction by Learning an Invariant Mapping , 2006, 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06).
[21] Binghui Wang,et al. On Certifying Robustness against Backdoor Attacks via Randomized Smoothing , 2020, ArXiv.
[22] Bo Li,et al. RAB: Provable Robustness Against Backdoor Attacks , 2020, 2023 IEEE Symposium on Security and Privacy (SP).
[23] Tom Goldstein,et al. Certified Defenses for Adversarial Patches , 2020, ICLR.
[24] Reza Shokri,et al. Bypassing Backdoor Detection Algorithms in Deep Learning , 2019, 2020 IEEE European Symposium on Security and Privacy (EuroS&P).
[25] Nitish Srivastava,et al. Multimodal learning with deep Boltzmann machines , 2012, J. Mach. Learn. Res..
[26] Brendan Dolan-Gavitt,et al. BadNets: Identifying Vulnerabilities in the Machine Learning Model Supply Chain , 2017, ArXiv.
[27] Jinyuan Jia,et al. Intrinsic Certified Robustness of Bagging against Data Poisoning Attacks , 2020, AAAI.
[28] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[29] Benjamin Edwards,et al. Detecting Backdoor Attacks on Deep Neural Networks by Activation Clustering , 2018, SafeAI@AAAI.
[30] Damith Chinthana Ranasinghe,et al. STRIP: a defence against trojan attacks on deep neural networks , 2019, ACSAC.
[31] Yang Zhang,et al. Dynamic Backdoor Attacks Against Machine Learning Models , 2020, 2022 IEEE 7th European Symposium on Security and Privacy (EuroS&P).
[32] Mark Chen,et al. Language Models are Few-Shot Learners , 2020, NeurIPS.
[33] Nicholas Carlini,et al. Poisoning and Backdooring Contrastive Learning , 2021, ICLR.
[34] Vitaly Shmatikov,et al. Blind Backdoors in Deep Learning Models , 2020, USENIX Security Symposium.
[35] Ilya Sutskever,et al. Learning Transferable Visual Models From Natural Language Supervision , 2021, ICML.
[36] Jerry Li,et al. Spectral Signatures in Backdoor Attacks , 2018, NeurIPS.
[37] Ting Wang,et al. Backdoor attacks against learning systems , 2017, 2017 IEEE Conference on Communications and Network Security (CNS).
[38] Brendan Dolan-Gavitt,et al. Fine-Pruning: Defending Against Backdooring Attacks on Deep Neural Networks , 2018, RAID.
[39] Alexei A. Efros,et al. Context Encoders: Feature Learning by Inpainting , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[40] Allan Jabri,et al. Learning Visual Features from Large Weakly Supervised Data , 2015, ECCV.
[41] Florian Tramèr,et al. SentiNet: Detecting Localized Universal Attacks Against Deep Learning Systems , 2018, 2020 IEEE Security and Privacy Workshops (SPW).
[42] Matthieu Guillaumin,et al. Food-101 - Mining Discriminative Components with Random Forests , 2014, ECCV.
[43] Yuxiao Dong,et al. GCC: Graph Contrastive Coding for Graph Neural Network Pre-Training , 2020, KDD.
[44] Ilya Sutskever,et al. Language Models are Unsupervised Multitask Learners , 2019 .
[45] Alexander Levine,et al. (De)Randomized Smoothing for Certifiable Defense against Patch Attacks , 2020, NeurIPS.
[46] Sencun Zhu,et al. Backdoor Embedding in Convolutional Neural Network Models via Invisible Perturbation , 2018, CODASPY.
[47] Yunfei Liu,et al. Reflection Backdoor: A Natural Backdoor Attack on Deep Neural Networks , 2020, ECCV.
[48] Jishen Zhao,et al. DeepInspect: A Black-box Trojan Detection and Mitigation Framework for Deep Neural Networks , 2019, IJCAI.
[49] Nikita Borisov,et al. Detecting AI Trojans Using Meta Neural Analysis , 2019, 2021 IEEE Symposium on Security and Privacy (SP).
[50] Ross B. Girshick,et al. Momentum Contrast for Unsupervised Visual Representation Learning , 2019, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[51] J. Leskovec,et al. Strategies for Pre-training Graph Neural Networks , 2019, ICLR.
[52] Jinyuan Jia,et al. Backdoor Attacks to Graph Neural Networks , 2020, SACMAT.
[53] Honglak Lee,et al. An Analysis of Single-Layer Networks in Unsupervised Feature Learning , 2011, AISTATS.
[54] Shouling Ji,et al. Graph Backdoor , 2020, USENIX Security Symposium.
[55] Johannes Stallkamp,et al. Man vs. computer: Benchmarking machine learning algorithms for traffic sign recognition , 2012, Neural Networks.
[56] Tudor Dumitras,et al. Poison Frogs! Targeted Clean-Label Poisoning Attacks on Neural Networks , 2018, NeurIPS.
[57] LE Lavage,et al. LBA , 2020, Catalysis from A to Z.
[58] Dawn Xiaodong Song,et al. Targeted Backdoor Attacks on Deep Learning Systems Using Data Poisoning , 2017, ArXiv.
[59] Alec Radford,et al. Improving Language Understanding by Generative Pre-Training , 2018 .
[60] Ben Y. Zhao,et al. Latent Backdoor Attacks on Deep Neural Networks , 2019, CCS.
[61] Xiangyu Zhang,et al. ABS: Scanning Neural Networks for Back-doors by Artificial Brain Stimulation , 2019, CCS.
[62] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[63] Michael Backes,et al. BadNL: Backdoor Attacks Against NLP Models , 2020, ArXiv.
[64] Yiming Yang,et al. XLNet: Generalized Autoregressive Pretraining for Language Understanding , 2019, NeurIPS.
[65] Yoshua Bengio,et al. Learning deep representations by mutual information estimation and maximization , 2018, ICLR.
[66] Yufeng Li,et al. A Backdoor Attack Against LSTM-Based Text Classification Systems , 2019, IEEE Access.
[67] Ting Wang,et al. Model-Reuse Attacks on Deep Learning Systems , 2018, CCS.
[68] Haixu Tang,et al. Demon in the Variant: Statistical Analysis of DNNs for Robust Backdoor Contamination Detection , 2019, USENIX Security Symposium.