Gradient Disaggregation: Breaking Privacy in Federated Learning by Reconstructing the User Participant Matrix
暂无分享,去创建一个
Gu-Yeon Wei | Michael Mitzenmacher | Vijay Janapa Reddi | David Brooks | Maximilian Lam | M. Mitzenmacher | Maximilian Lam | Gu-Yeon Wei | D. Brooks | V. Reddi
[1] Blaise Agüera y Arcas,et al. Communication-Efficient Learning of Deep Networks from Decentralized Data , 2016, AISTATS.
[2] Song Han,et al. Deep Leakage from Gradients , 2019, NeurIPS.
[3] Vitaly Shmatikov,et al. Membership Inference Attacks Against Machine Learning Models , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[4] A. Salman Avestimehr,et al. Turbo-Aggregate: Breaking the Quadratic Aggregation Barrier in Secure Federated Learning , 2020, IEEE Journal on Selected Areas in Information Theory.
[5] Yanzhao Wu,et al. A Framework for Evaluating Gradient Leakage Attacks in Federated Learning , 2020, ArXiv.
[6] Jakub Konecný,et al. Federated Optimization: Distributed Optimization Beyond the Datacenter , 2015, ArXiv.
[7] Hubert Eichner,et al. Towards Federated Learning at Scale: System Design , 2019, MLSys.
[8] Vitaly Shmatikov,et al. Exploiting Unintended Feature Leakage in Collaborative Learning , 2018, 2019 IEEE Symposium on Security and Privacy (SP).
[9] Hubert Eichner,et al. Federated Learning for Mobile Keyboard Prediction , 2018, ArXiv.
[10] Bo Chen,et al. MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications , 2017, ArXiv.
[11] Han Yu,et al. Threats to Federated Learning: A Survey , 2020, ArXiv.
[12] Anit Kumar Sahu,et al. Federated Learning: Challenges, Methods, and Future Directions , 2019, IEEE Signal Processing Magazine.
[13] Song Han,et al. Learning both Weights and Connections for Efficient Neural Network , 2015, NIPS.
[14] Vitaly Shmatikov,et al. How To Backdoor Federated Learning , 2018, AISTATS.
[15] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[16] Giuseppe Ateniese,et al. Deep Models Under the GAN: Information Leakage from Collaborative Deep Learning , 2017, CCS.
[17] Lars Kai Hansen,et al. What can we learn from gradients? , 2020, ArXiv.
[18] Kartik Sreenivasan,et al. Attack of the Tails: Yes, You Really Can Backdoor Federated Learning , 2020, NeurIPS.
[19] Rachid Guerraoui,et al. Machine Learning with Adversaries: Byzantine Tolerant Gradient Descent , 2017, NIPS.
[20] Matthias Hein,et al. Matrix factorization with binary components , 2013, NIPS.
[21] Song Han,et al. Deep Compression: Compressing Deep Neural Network with Pruning, Trained Quantization and Huffman Coding , 2015, ICLR.
[22] Peter Richtárik,et al. Federated Learning: Strategies for Improving Communication Efficiency , 2016, ArXiv.
[23] Marwan Mattar,et al. Labeled Faces in the Wild: A Database forStudying Face Recognition in Unconstrained Environments , 2008 .
[24] Ivan Beschastnikh,et al. Mitigating Sybils in Federated Learning Poisoning , 2018, ArXiv.
[25] Devavrat Shah,et al. I Know What You Bought At Chipotle for $9.81 by Solving A Linear Inverse Problem , 2020, Proc. ACM Meas. Anal. Comput. Syst..
[26] Zhifei Zhang,et al. Analyzing User-Level Privacy Attack Against Federated Learning , 2020, IEEE Journal on Selected Areas in Communications.
[27] David A. Wagner,et al. Obfuscated Gradients Give a False Sense of Security: Circumventing Defenses to Adversarial Examples , 2018, ICML.
[28] Richard Nock,et al. Advances and Open Problems in Federated Learning , 2021, Found. Trends Mach. Learn..
[29] Sarvar Patel,et al. Practical Secure Aggregation for Privacy-Preserving Machine Learning , 2017, IACR Cryptol. ePrint Arch..
[30] Michael Moeller,et al. Inverting Gradients - How easy is it to break privacy in federated learning? , 2020, NeurIPS.