Intrinsic Certified Robustness of Bagging against Data Poisoning Attacks
暂无分享,去创建一个
[1] Shie Mannor,et al. Robust Logistic Regression and Classification , 2014, NIPS.
[2] Fabio Roli,et al. Towards Poisoning of Deep Learning Algorithms with Back-gradient Optimization , 2017, AISec@CCS.
[3] Wen-Chuan Lee,et al. Trojaning Attack on Neural Networks , 2018, NDSS.
[4] Jinyuan Jia,et al. Local Model Poisoning Attacks to Byzantine-Robust Federated Learning , 2019, USENIX Security Symposium.
[5] Percy Liang,et al. Stronger data poisoning attacks break data sanitization defenses , 2018, Machine Learning.
[6] Vitaly Shmatikov,et al. How To Backdoor Federated Learning , 2018, AISTATS.
[7] E. S. Pearson,et al. THE USE OF CONFIDENCE OR FIDUCIAL LIMITS ILLUSTRATED IN THE CASE OF THE BINOMIAL , 1934 .
[8] Ling Huang,et al. ANTIDOTE: understanding and defending against poisoning of anomaly detectors , 2009, IMC '09.
[9] Ying Cai,et al. Fake Co-visitation Injection Attacks to Recommender Systems , 2017, NDSS.
[10] Yiran Chen,et al. Generative Poisoning Attack Method Against Neural Networks , 2017, ArXiv.
[11] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[12] Claudia Eckert,et al. Is Feature Selection Secure against Training Data Poisoning? , 2015, ICML.
[13] Xiaojin Zhu,et al. The Security of Latent Dirichlet Allocation , 2015, AISTATS.
[14] Jinyuan Jia,et al. Certified Robustness of Community Detection against Adversarial Structural Perturbation via Randomized Smoothing , 2020, WWW.
[15] Binghui Wang,et al. On Certifying Robustness against Backdoor Attacks via Randomized Smoothing , 2020, ArXiv.
[16] Yevgeniy Vorobeychik,et al. Data Poisoning Attacks on Factorization-Based Collaborative Filtering , 2016, NIPS.
[17] Binghui Wang,et al. Attacking Graph-based Classification via Manipulating the Graph Structure , 2019, CCS.
[18] Brendan Dolan-Gavitt,et al. BadNets: Identifying Vulnerabilities in the Machine Learning Model Supply Chain , 2017, ArXiv.
[19] Justin Hsu,et al. Data Poisoning against Differentially-Private Learners: Attacks and Defenses , 2019, IJCAI.
[20] J. Z. Kolter,et al. Certified Robustness to Label-Flipping Attacks via Randomized Smoothing , 2020, ICML.
[21] Xiaoyu Cao,et al. Mitigating Evasion Attacks to Deep Neural Networks via Region-based Classification , 2017, ACSAC.
[22] Stephan Günnemann,et al. Adversarial Attacks on Neural Networks for Graph Data , 2018, KDD.
[23] Prateek Mittal,et al. Analyzing Federated Learning through an Adversarial Lens , 2018, ICML.
[24] Alexander Levine,et al. Deep Partition Aggregation: Provable Defense against General Poisoning Attacks , 2020, ICLR.
[25] Jerry Li,et al. Spectral Signatures in Backdoor Attacks , 2018, NeurIPS.
[26] Tudor Dumitras,et al. When Does Machine Learning FAIL? Generalized Transferability for Evasion and Poisoning Attacks , 2018, USENIX Security Symposium.
[27] Binghui Wang,et al. Certified Robustness for Top-k Predictions against Adversarial Perturbations via Randomized Smoothing , 2019, ICLR.
[28] Xiaojin Zhu,et al. Using Machine Teaching to Identify Optimal Training-Set Attacks on Machine Learners , 2015, AAAI.
[29] Blaine Nelson,et al. Poisoning Attacks against Support Vector Machines , 2012, ICML.
[30] E. S. Pearson,et al. On the Problem of the Most Efficient Tests of Statistical Hypotheses , 1933 .
[31] Blaine Nelson,et al. Exploiting Machine Learning to Subvert Your Spam Filter , 2008, LEET.
[32] Jia Liu,et al. Poisoning Attacks to Graph-Based Recommender Systems , 2018, ACSAC.
[33] Blaine Nelson,et al. The security of machine learning , 2010, Machine Learning.
[34] Fabio Roli,et al. Poisoning behavioral malware clustering , 2014, AISec '14.
[35] Jia Liu,et al. Influence Function based Data Poisoning Attacks to Top-N Recommender Systems , 2020, WWW.
[36] Jinyuan Jia,et al. Backdoor Attacks to Graph Neural Networks , 2020, SACMAT.
[37] Siddharth Garg,et al. BadNets: Evaluating Backdooring Attacks on Deep Neural Networks , 2019, IEEE Access.
[38] Fabio Roli,et al. Bagging Classifiers for Fighting Poisoning Attacks in Adversarial Classification Tasks , 2011, MCS.
[39] Fabio Roli,et al. Is data clustering in adversarial settings secure? , 2013, AISec.
[40] Tudor Dumitras,et al. Poison Frogs! Targeted Clean-Label Poisoning Attacks on Neural Networks , 2018, NeurIPS.
[41] Susmita Sur-Kolay,et al. Systematic Poisoning Attacks on and Defenses for Machine Learning in Healthcare , 2015, IEEE Journal of Biomedical and Health Informatics.
[42] Leo Breiman,et al. Bagging Predictors , 1996, Machine Learning.
[43] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[44] Percy Liang,et al. Certified Defenses for Data Poisoning Attacks , 2017, NIPS.
[45] Tom Goldstein,et al. Transferable Clean-Label Poisoning Attacks on Deep Neural Nets , 2019, ICML.
[46] David A. Wagner,et al. Towards Evaluating the Robustness of Neural Networks , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[47] Salvatore J. Stolfo,et al. Casting out Demons: Sanitizing Training Data for Anomaly Sensors , 2008, 2008 IEEE Symposium on Security and Privacy (sp 2008).
[48] Ian Goodfellow,et al. Deep Learning with Differential Privacy , 2016, CCS.
[49] Chang Liu,et al. Manipulating Machine Learning: Poisoning Attacks and Countermeasures for Regression Learning , 2018, 2018 IEEE Symposium on Security and Privacy (SP).
[50] J. Zico Kolter,et al. Certified Adversarial Robustness via Randomized Smoothing , 2019, ICML.