A Survey of Adversarial Machine Learning in Cyber Warfare
暂无分享,去创建一个
[1] Ananthram Swami,et al. Crafting adversarial input sequences for recurrent neural networks , 2016, MILCOM 2016 - 2016 IEEE Military Communications Conference.
[2] Fabio Roli,et al. Poisoning attacks to compromise face templates , 2013, 2013 International Conference on Biometrics (ICB).
[3] Tobias Scheffer,et al. Bayesian Games for Adversarial Regression Problems , 2013, ICML.
[4] Yoshua Bengio,et al. Generative Adversarial Nets , 2014, NIPS.
[5] David B. Skillicorn,et al. Hiding clusters in adversarial settings , 2008, 2008 IEEE International Conference on Intelligence and Security Informatics.
[6] Martín Abadi,et al. Semi-supervised Knowledge Transfer for Deep Learning from Private Training Data , 2016, ICLR.
[7] Ian Goodfellow,et al. Deep Learning with Differential Privacy , 2016, CCS.
[8] Patrick D. McDaniel,et al. Extending Defensive Distillation , 2017, ArXiv.
[9] Nina Narodytska,et al. Simple Black-Box Adversarial Attacks on Deep Neural Networks , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).
[10] Patrick D. McDaniel,et al. Transferability in Machine Learning: from Phenomena to Black-Box Attacks using Adversarial Samples , 2016, ArXiv.
[11] Dan Boneh,et al. Ensemble Adversarial Training: Attacks and Defenses , 2017, ICLR.
[12] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[13] Christopher Meek,et al. Adversarial learning , 2005, KDD '05.
[14] Fabio Roli,et al. Adversarial Pattern Classification Using Multiple Classifiers and Randomisation , 2008, SSPR/SPR.
[15] Claudia Eckert,et al. Support vector machines under adversarial label contamination , 2015, Neurocomputing.
[16] Ananthram Swami,et al. Practical Black-Box Attacks against Machine Learning , 2016, AsiaCCS.
[17] Payman Mohassel,et al. SecureML: A System for Scalable Privacy-Preserving Machine Learning , 2017, 2017 IEEE Symposium on Security and Privacy (SP).
[18] Ananthram Swami,et al. The Limitations of Deep Learning in Adversarial Settings , 2015, 2016 IEEE European Symposium on Security and Privacy (EuroS&P).
[19] Fabio Roli,et al. Multiple Classifier Systems under Attack , 2010, MCS.
[20] Fabio Roli,et al. Security Evaluation of Pattern Classifiers under Attack , 2014, IEEE Transactions on Knowledge and Data Engineering.
[21] Xirong Li,et al. Deep Text Classification Can be Fooled , 2017, IJCAI.
[22] Claudia Eckert,et al. Is Feature Selection Secure against Training Data Poisoning? , 2015, ICML.
[23] Dan Boneh,et al. The Space of Transferable Adversarial Examples , 2017, ArXiv.
[24] Dawn Xiaodong Song,et al. Adversarial Examples for Generative Models , 2017, 2018 IEEE Security and Privacy Workshops (SPW).
[25] Samy Bengio,et al. Adversarial Machine Learning at Scale , 2016, ICLR.
[26] Sebastian Nowozin,et al. Oblivious Multi-Party Machine Learning on Trusted Processors , 2016, USENIX Security Symposium.
[27] David A. Wagner,et al. Towards Evaluating the Robustness of Neural Networks , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[28] Michael P. Wellman,et al. Towards the Science of Security and Privacy in Machine Learning , 2016, ArXiv.
[29] Claudia Eckert,et al. Adversarial Label Flips Attack on Support Vector Machines , 2012, ECAI.
[30] Ananthram Swami,et al. Distillation as a Defense to Adversarial Perturbations Against Deep Neural Networks , 2015, 2016 IEEE Symposium on Security and Privacy (SP).
[31] Dawn Xiaodong Song,et al. Delving into adversarial attacks on deep policies , 2017, ICLR.
[32] Samy Bengio,et al. Adversarial examples in the physical world , 2016, ICLR.
[33] George Danezis,et al. Machine Learning as an Adversarial Service: Learning Black-Box Adversarial Examples , 2017, ArXiv.
[34] Blaine Nelson,et al. The security of machine learning , 2010, Machine Learning.
[35] Fabio Roli,et al. Evasion Attacks against Machine Learning at Test Time , 2013, ECML/PKDD.
[36] Blaine Nelson,et al. Adversarial machine learning , 2019, AISec '11.
[37] Brendan Dolan-Gavitt,et al. BadNets: Identifying Vulnerabilities in the Machine Learning Model Supply Chain , 2017, ArXiv.
[38] Alan L. Yuille,et al. Adversarial Examples for Semantic Segmentation and Object Detection , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[39] Somesh Jha,et al. Model Inversion Attacks that Exploit Confidence Information and Basic Countermeasures , 2015, CCS.
[40] Daniel Cullina,et al. Enhancing robustness of machine learning systems via data transformations , 2017, 2018 52nd Annual Conference on Information Sciences and Systems (CISS).
[41] Khorshidpour Zeinab,et al. Learning a Secure Classifier against Evasion Attack , 2016 .
[42] Blaine Nelson,et al. Support Vector Machines Under Adversarial Label Noise , 2011, ACML.
[43] Sandy H. Huang,et al. Adversarial Attacks on Neural Network Policies , 2017, ICLR.
[44] Yanjun Qi,et al. Feature Squeezing: Detecting Adversarial Examples in Deep Neural Networks , 2017, NDSS.
[45] Yanjun Qi,et al. Automatically Evading Classifiers: A Case Study on PDF Malware Classifiers , 2016, NDSS.
[46] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[47] David A. Wagner,et al. Defensive Distillation is Not Robust to Adversarial Examples , 2016, ArXiv.
[48] Pavel Laskov,et al. Practical Evasion of a Learning-Based Classifier: A Case Study , 2014, 2014 IEEE Symposium on Security and Privacy.
[49] Geoffrey E. Hinton,et al. Distilling the Knowledge in a Neural Network , 2015, ArXiv.
[50] Fabio Roli,et al. Is data clustering in adversarial settings secure? , 2013, AISec.
[51] Fabio Roli,et al. Poisoning Complete-Linkage Hierarchical Clustering , 2014, S+SSPR.
[52] Navdeep Jaitly,et al. Adversarial Autoencoders , 2015, ArXiv.
[53] D. Vijay Rao,et al. A Game Theoretic Approach to Modelling Jamming Attacks in Delay Tolerant Networks , 2017 .
[54] Blaine Nelson,et al. Poisoning Attacks against Support Vector Machines , 2012, ICML.
[55] Dawn Xiaodong Song,et al. Delving into Transferable Adversarial Examples and Black-box Attacks , 2016, ICLR.
[56] Brent Lagesse,et al. Analysis of Causative Attacks against SVMs Learning from Data Streams , 2017, IWSPA@CODASPY.
[57] Li Guo,et al. Adversarial Reinforcement Learning for Chinese Text Summarization , 2018, ICCS.
[58] Ming-Yu Liu,et al. Tactics of Adversarial Attack on Deep Reinforcement Learning Agents , 2017, IJCAI.
[59] Andrew M. Dai,et al. Adversarial Training Methods for Semi-Supervised Text Classification , 2016, ICLR.
[60] Logan Engstrom,et al. Synthesizing Robust Adversarial Examples , 2017, ICML.
[61] Patrick P. K. Chan,et al. Adversarial Feature Selection Against Evasion Attacks , 2016, IEEE Transactions on Cybernetics.
[62] Fan Zhang,et al. Stealing Machine Learning Models via Prediction APIs , 2016, USENIX Security Symposium.
[63] Ying Tan,et al. Generating Adversarial Malware Examples for Black-Box Attacks Based on GAN , 2017, DMBD.
[64] Michael Naehrig,et al. CryptoNets: applying neural networks to encrypted data with high throughput and accuracy , 2016, ICML 2016.
[65] D. Vijay Rao,et al. Game Theory-Based Defense Mechanisms of Cyber Warfare , 2018 .
[66] Farinaz Koushanfar,et al. DeepSecure: Scalable Provably-Secure Deep Learning , 2017, 2018 55th ACM/ESDA/IEEE Design Automation Conference (DAC).
[67] David A. Forsyth,et al. SafetyNet: Detecting and Rejecting Adversarial Examples Robustly , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[68] D. Vijay Rao,et al. A Game Theoretic Software Test-bed for Cyber Security Analysis of Critical Infrastructure , 2017 .
[69] Yongdong Zhang,et al. APE-GAN: Adversarial Perturbation Elimination with GAN , 2017, ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[70] Vitaly Shmatikov,et al. Membership Inference Attacks Against Machine Learning Models , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[71] Arslan Munir,et al. Vulnerability of Deep Reinforcement Learning to Policy Induction Attacks , 2017, MLDM.
[72] Vitaly Shmatikov,et al. Privacy-preserving deep learning , 2015, 2015 53rd Annual Allerton Conference on Communication, Control, and Computing (Allerton).
[73] Seyed-Mohsen Moosavi-Dezfooli,et al. DeepFool: A Simple and Accurate Method to Fool Deep Neural Networks , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[74] Marius Kloft,et al. Online Anomaly Detection under Adversarial Impact , 2010, AISTATS.
[75] Fabio Roli,et al. Towards Poisoning of Deep Learning Algorithms with Back-gradient Optimization , 2017, AISec@CCS.
[76] Fabio Roli,et al. Poisoning behavioral malware clustering , 2014, AISec '14.
[77] Susmita Sur-Kolay,et al. Systematic Poisoning Attacks on and Defenses for Machine Learning in Healthcare , 2015, IEEE Journal of Biomedical and Health Informatics.
[78] Fabio Roli,et al. Yes, Machine Learning Can Be More Secure! A Case Study on Android Malware Detection , 2017, IEEE Transactions on Dependable and Secure Computing.