暂无分享,去创建一个
Tianjian Chen | Yang Liu | Suyi Li | Wei Wang | Yong Cheng
[1] Xiang Li,et al. On the Convergence of FedAvg on Non-IID Data , 2019, ICLR.
[2] Blaise Agüera y Arcas,et al. Federated Learning of Deep Networks using Model Averaging , 2016, ArXiv.
[3] Zoubin Ghahramani,et al. Unifying linear dimensionality reduction , 2014, 1406.0873.
[4] Martin Jaggi,et al. Sparsified SGD with Memory , 2018, NeurIPS.
[5] Ian Goodfellow,et al. Deep Learning with Differential Privacy , 2016, CCS.
[6] Zhiwei Steven Wu,et al. Distributed Training with Heterogeneous Data: Bridging Median- and Mean-Based Algorithms , 2019, NeurIPS.
[7] Kannan Ramchandran,et al. Byzantine-Robust Distributed Learning: Towards Optimal Statistical Rates , 2018, ICML.
[8] Spyridon Bakas,et al. Multi-Institutional Deep Learning Modeling Without Sharing Patient Data: A Feasibility Study on Brain Tumor Segmentation , 2018, BrainLes@MICCAI.
[9] Qiang Yang,et al. Federated Machine Learning , 2019, ACM Trans. Intell. Syst. Technol..
[10] Qing Ling,et al. RSA: Byzantine-Robust Stochastic Aggregation Methods for Distributed Learning from Heterogeneous Datasets , 2018, AAAI.
[11] Ameet Talwalkar,et al. Federated Multi-Task Learning , 2017, NIPS.
[12] Sebastian Nowozin,et al. Oblivious Multi-Party Machine Learning on Trusted Processors , 2016, USENIX Security Symposium.
[13] Anit Kumar Sahu,et al. Federated Optimization in Heterogeneous Networks , 2018, MLSys.
[14] Hubert Eichner,et al. Towards Federated Learning at Scale: System Design , 2019, SysML.
[15] Tetsuro Morimura,et al. Weight Features for Predicting Future Model Performance of Deep Neural Networks , 2016, IJCAI.
[16] Shenghuo Zhu,et al. Parallel Restarted SGD with Faster Convergence and Less Communication: Demystifying Why Model Averaging Works for Deep Learning , 2018, AAAI.
[17] Vitaly Shmatikov,et al. How To Backdoor Federated Learning , 2018, AISTATS.
[18] Prateek Mittal,et al. Analyzing Federated Learning through an Adversarial Lens , 2018, ICML.
[19] Tassilo Klein,et al. Differentially Private Federated Learning: A Client Level Perspective , 2017, ArXiv.
[20] Ying-Chang Liang,et al. Incentive Design for Efficient Federated Learning in Mobile Networks: A Contract Theory Approach , 2019, 2019 IEEE VTS Asia Pacific Wireless Communications Symposium (APWCS).
[21] Peter Richtárik,et al. Federated Learning: Strategies for Improving Communication Efficiency , 2016, ArXiv.
[22] Yoshua Bengio,et al. Gradient-based learning applied to document recognition , 1998, Proc. IEEE.
[23] Le Trieu Phong,et al. Privacy-Preserving Deep Learning via Weight Transmission , 2018, IEEE Transactions on Information Forensics and Security.
[24] Sebastian Caldas,et al. LEAF: A Benchmark for Federated Settings , 2018, ArXiv.
[25] Blaise Agüera y Arcas,et al. Communication-Efficient Learning of Deep Networks from Decentralized Data , 2016, AISTATS.
[26] Rachid Guerraoui,et al. Machine Learning with Adversaries: Byzantine Tolerant Gradient Descent , 2017, NIPS.
[27] VARUN CHANDOLA,et al. Anomaly detection: A survey , 2009, CSUR.
[28] Prateek Saxena,et al. Auror: defending against poisoning attacks in collaborative deep learning systems , 2016, ACSAC.
[29] Sarvar Patel,et al. Practical Secure Aggregation for Federated Learning on User-Held Data , 2016, ArXiv.
[30] Lili Su,et al. Distributed Statistical Machine Learning in Adversarial Settings: Byzantine Gradient Descent , 2017, Abstracts of the 2018 ACM International Conference on Measurement and Modeling of Computer Systems.
[31] Takehisa Yairi,et al. Anomaly Detection Using Autoencoders with Nonlinear Dimensionality Reduction , 2014, MLSDA'14.