暂无分享,去创建一个
Shucheng Yu | Christian Makaya | Zhuosheng Zhang | Jiarui Li | C. Makaya | Shucheng Yu | Zhuosheng Zhang | Jiarui Li
[1] John R. Douceur,et al. The Sybil Attack , 2002, IPTPS.
[2] Dawn Xiaodong Song,et al. Targeted Backdoor Attacks on Deep Learning Systems Using Data Poisoning , 2017, ArXiv.
[3] Shafi Goldwasser,et al. Machine Learning Classification over Encrypted Data , 2015, NDSS.
[4] Claude Castelluccia,et al. I Have a DREAM! (DiffeRentially privatE smArt Metering) , 2011, Information Hiding.
[5] Brendan Dolan-Gavitt,et al. BadNets: Identifying Vulnerabilities in the Machine Learning Model Supply Chain , 2017, ArXiv.
[6] Wen-Chuan Lee,et al. Trojaning Attack on Neural Networks , 2018, NDSS.
[7] Payman Mohassel,et al. SecureML: A System for Scalable Privacy-Preserving Machine Learning , 2017, 2017 IEEE Symposium on Security and Privacy (SP).
[8] Ivan Beschastnikh,et al. Mitigating Sybils in Federated Learning Poisoning , 2018, ArXiv.
[9] Rachid Guerraoui,et al. Machine Learning with Adversaries: Byzantine Tolerant Gradient Descent , 2017, NIPS.
[10] Elaine Shi,et al. Privacy-Preserving Stream Aggregation with Fault Tolerance , 2012, Financial Cryptography.
[11] Peter Rindal,et al. ABY3: A Mixed Protocol Framework for Machine Learning , 2018, IACR Cryptol. ePrint Arch..
[12] Kenneth T. Co,et al. Byzantine-Robust Federated Machine Learning through Adaptive Model Averaging , 2019, ArXiv.
[13] Anantha Chandrakasan,et al. Gazelle: A Low Latency Framework for Secure Neural Network Inference , 2018, IACR Cryptol. ePrint Arch..
[14] Vitaly Shmatikov,et al. How To Backdoor Federated Learning , 2018, AISTATS.
[15] Sameer Wagh,et al. SecureNN: Efficient and Private Neural Network Training , 2018, IACR Cryptol. ePrint Arch..
[16] Whitfield Diffie,et al. New Directions in Cryptography , 1976, IEEE Trans. Inf. Theory.
[17] Blaine Nelson,et al. Poisoning Attacks against Support Vector Machines , 2012, ICML.
[18] Sarvar Patel,et al. Practical Secure Aggregation for Privacy-Preserving Machine Learning , 2017, IACR Cryptol. ePrint Arch..
[19] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[20] Prateek Mittal,et al. Analyzing Federated Learning through an Adversarial Lens , 2018, ICML.
[21] Elaine Shi,et al. Privacy-Preserving Aggregation of Time-Series Data , 2011, NDSS.
[22] Yao Lu,et al. Oblivious Neural Network Predictions via MiniONN Transformations , 2017, IACR Cryptol. ePrint Arch..
[23] Farinaz Koushanfar,et al. Chameleon: A Hybrid Secure Computation Framework for Machine Learning Applications , 2018, IACR Cryptol. ePrint Arch..
[24] Peter Richtárik,et al. Federated Learning: Strategies for Improving Communication Efficiency , 2016, ArXiv.
[25] Raymond H. Myers,et al. Probability and Statistics for Engineers and Scientists. , 1973 .
[26] Stratis Ioannidis,et al. Privacy-Preserving Ridge Regression on Hundreds of Millions of Records , 2013, 2013 IEEE Symposium on Security and Privacy.
[27] J. Doug Tygar,et al. Adversarial machine learning , 2019, AISec '11.
[28] Tassilo Klein,et al. Differentially Private Federated Learning: A Client Level Perspective , 2017, ArXiv.
[29] Bo Li,et al. DBA: Distributed Backdoor Attacks against Federated Learning , 2020, ICLR.
[30] Prateek Saxena,et al. Auror: defending against poisoning attacks in collaborative deep learning systems , 2016, ACSAC.
[31] Ian Goodfellow,et al. Deep Learning with Differential Privacy , 2016, CCS.
[32] Mohammad Al-Rubaie,et al. Privacy-Preserving Machine Learning: Threats and Solutions , 2018, IEEE Security & Privacy.
[33] Blaise Agüera y Arcas,et al. Communication-Efficient Learning of Deep Networks from Decentralized Data , 2016, AISTATS.
[34] Li Xiong,et al. A Comprehensive Comparison of Multiparty Secure Additions with Differential Privacy , 2017, IEEE Transactions on Dependable and Secure Computing.
[35] Pritish Narayanan,et al. Deep Learning with Limited Numerical Precision , 2015, ICML.