Influence Function based Data Poisoning Attacks to Top-N Recommender Systems
暂无分享,去创建一个
[1] Wen-Chuan Lee,et al. Trojaning Attack on Neural Networks , 2018, NDSS.
[2] Kannan Ramchandran,et al. Byzantine-Robust Distributed Learning: Towards Optimal Statistical Rates , 2018, ICML.
[3] Bhaskar Mehta,et al. Attack resistant collaborative filtering , 2008, SIGIR '08.
[4] Konstantina Christakopoulou,et al. Adversarial attacks on an oblivious recommender , 2019, RecSys.
[5] Chang Liu,et al. Manipulating Machine Learning: Poisoning Attacks and Countermeasures for Regression Learning , 2018, 2018 IEEE Symposium on Security and Privacy (SP).
[6] Percy Liang,et al. On the Accuracy of Influence Functions for Measuring Group Effects , 2019, NeurIPS.
[7] Ling Huang,et al. ANTIDOTE: understanding and defending against poisoning of anomaly detectors , 2009, IMC '09.
[8] Hanif D. Sherali,et al. Linear Programming and Network Flows: Bazaraa/Linear , 2009 .
[9] Blaine Nelson,et al. Poisoning Attacks against Support Vector Machines , 2012, ICML.
[10] Wolfgang Nejdl,et al. Preventing shilling attacks in online recommender systems , 2005, WIDM '05.
[11] Ying Cai,et al. Fake Co-visitation Injection Attacks to Recommender Systems , 2017, NDSS.
[12] Zhengyuan Zhu,et al. Private and communication-efficient edge learning: a sparse differential gaussian-masking distributed SGD approach , 2020, MobiHoc.
[13] Nick Feamster,et al. Take This Personally: Pollution Attacks on Personalized Services , 2013, USENIX Security Symposium.
[14] Reza Shokri,et al. Comprehensive Privacy Analysis of Deep Learning: Stand-alone and Federated Learning under Passive and Active White-box Inference Attacks , 2018, ArXiv.
[15] Jia Liu,et al. Byzantine-Resilient Stochastic Gradient Descent for Distributed Learning: A Lipschitz-Inspired Coordinate-wise Median Approach , 2019, 2019 IEEE 58th Conference on Decision and Control (CDC).
[16] Binghui Wang,et al. Stealing Hyperparameters in Machine Learning , 2018, 2018 IEEE Symposium on Security and Privacy (SP).
[17] John Riedl,et al. Shilling recommender systems for fun and profit , 2004, WWW '04.
[18] Dawn Xiaodong Song,et al. Targeted Backdoor Attacks on Deep Learning Systems Using Data Poisoning , 2017, ArXiv.
[19] Yevgeniy Vorobeychik,et al. Data Poisoning Attacks on Factorization-Based Collaborative Filtering , 2016, NIPS.
[20] Jure Leskovec,et al. Supervised random walks: predicting and recommending links in social networks , 2010, WSDM '11.
[21] Bamshad Mobasher,et al. Towards Trustworthy Recommender Systems : An Analysis of Attack Models and Algorithm Robustness , 2007 .
[22] S. Weisberg,et al. Characterizations of an Empirical Influence Function for Detecting Influential Cases in Regression , 1980 .
[23] Yehuda Koren,et al. Matrix Factorization Techniques for Recommender Systems , 2009, Computer.
[24] David C. Wilson,et al. Attacking item-based recommender systems with power items , 2014, RecSys '14.
[25] David C. Wilson,et al. When power users attack: assessing impacts in collaborative recommender systems , 2013, RecSys.
[26] Percy Liang,et al. Understanding Black-box Predictions via Influence Functions , 2017, ICML.
[27] Jia Liu,et al. Poisoning Attacks to Graph-Based Recommender Systems , 2018, ACSAC.
[28] François Fouss,et al. Random-Walk Computation of Similarities between Nodes of a Graph with Application to Collaborative Recommendation , 2007, IEEE Transactions on Knowledge and Data Engineering.
[29] Bo Li,et al. Data Dropout: Optimizing Training Data for Convolutional Neural Networks , 2018, 2018 IEEE 30th International Conference on Tools with Artificial Intelligence (ICTAI).
[30] Amir Houmansadr,et al. Comprehensive Privacy Analysis of Deep Learning: Passive and Active White-box Inference Attacks against Centralized and Federated Learning , 2018, 2019 IEEE Symposium on Security and Privacy (SP).
[31] Claudia Eckert,et al. Adversarial Label Flips Attack on Support Vector Machines , 2012, ECAI.
[32] M. L. Fisher,et al. An analysis of approximations for maximizing submodular set functions—I , 1978, Math. Program..
[33] Jon Kleinberg,et al. Maximizing the spread of influence through a social network , 2003, KDD '03.
[34] Vitaly Shmatikov,et al. Membership Inference Attacks Against Machine Learning Models , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[35] Blaine Nelson,et al. Exploiting Machine Learning to Subvert Your Spam Filter , 2008, LEET.
[36] Brendan Dolan-Gavitt,et al. BadNets: Identifying Vulnerabilities in the Machine Learning Model Supply Chain , 2017, ArXiv.
[37] Tudor Dumitras,et al. Poison Frogs! Targeted Clean-Label Poisoning Attacks on Neural Networks , 2018, NeurIPS.
[38] Jinyuan Jia,et al. Local Model Poisoning Attacks to Byzantine-Robust Federated Learning , 2019, USENIX Security Symposium.
[39] Binghui Wang,et al. Attacking Graph-based Classification via Manipulating the Graph Structure , 2019, CCS.
[40] Stephan Günnemann,et al. Adversarial Attacks on Neural Networks for Graph Data , 2018, KDD.
[41] Claudia Eckert,et al. Is Feature Selection Secure against Training Data Poisoning? , 2015, ICML.
[42] Hanif D. Sherali,et al. Linear programming and network flows (2nd ed.) , 1990 .
[43] Antonio Torralba,et al. Are all training examples equally valuable? , 2013, ArXiv.
[44] Robin Burke,et al. Effective Attack Models for Shilling Item-Based Collaborative Filtering Systems , 2005 .