暂无分享,去创建一个
[1] Hubert Eichner,et al. Towards Federated Learning at Scale: System Design , 2019, SysML.
[2] Sergey Ioffe,et al. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.
[3] Sreeram Kannan,et al. Improving Federated Learning Personalization via Model Agnostic Meta Learning , 2019, ArXiv.
[4] Peter Richtárik,et al. Federated Learning: Strategies for Improving Communication Efficiency , 2016, ArXiv.
[5] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[6] Yoshua Bengio,et al. Gradient-based learning applied to document recognition , 1998, Proc. IEEE.
[7] Junpu Wang,et al. FedMD: Heterogenous Federated Learning via Model Distillation , 2019, ArXiv.
[8] Ruslan Salakhutdinov,et al. Think Locally, Act Globally: Federated Learning with Local and Global Representations , 2020, ArXiv.
[9] Ying-Chang Liang,et al. Federated Learning in Mobile Edge Networks: A Comprehensive Survey , 2020, IEEE Communications Surveys & Tutorials.
[10] Vitaly Shmatikov,et al. Exploiting Unintended Feature Leakage in Collaborative Learning , 2018, 2019 IEEE Symposium on Security and Privacy (SP).
[11] Richard Socher,et al. Pointer Sentinel Mixture Models , 2016, ICLR.
[12] Takayuki Nishio,et al. Client Selection for Federated Learning with Heterogeneous Resources in Mobile Edge , 2018, ICC 2019 - 2019 IEEE International Conference on Communications (ICC).
[13] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[14] Quoc V. Le,et al. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks , 2019, ICML.
[15] Seyit Camtepe,et al. SplitFed: When Federated Learning Meets Split Learning , 2020, AAAI.
[16] Andrea Vedaldi,et al. Instance Normalization: The Missing Ingredient for Fast Stylization , 2016, ArXiv.
[17] Nikos Komodakis,et al. Wide Residual Networks , 2016, BMVC.
[18] Geoffrey E. Hinton,et al. Layer Normalization , 2016, ArXiv.
[19] Ameet Talwalkar,et al. Federated Multi-Task Learning , 2017, NIPS.
[20] Dan Alistarh,et al. QSGD: Communication-Optimal Stochastic Gradient Descent, with Applications to Training Neural Networks , 2016, 1610.02132.
[21] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[22] Hubert Eichner,et al. Federated Learning for Mobile Keyboard Prediction , 2018, ArXiv.
[23] Jian Sun,et al. Identity Mappings in Deep Residual Networks , 2016, ECCV.
[24] Anit Kumar Sahu,et al. Federated Learning: Challenges, Methods, and Future Directions , 2019, IEEE Signal Processing Magazine.
[25] Hubert Eichner,et al. Federated Evaluation of On-device Personalization , 2019, ArXiv.
[26] Kaiming He,et al. Group Normalization , 2018, International Journal of Computer Vision.
[27] Constance Beguier,et al. Siloed Federated Learning for Multi-Centric Histopathology Datasets , 2020, DART/DCL@MICCAI.
[28] Torsten Hoefler,et al. Demystifying Parallel and Distributed Deep Learning: An In-Depth Concurrency Analysis. , 2018 .
[29] Han Yu,et al. Threats to Federated Learning: A Survey , 2020, ArXiv.
[30] Yishay Mansour,et al. Three Approaches for Personalization with Applications to Federated Learning , 2020, ArXiv.
[31] Blaise Agüera y Arcas,et al. Communication-Efficient Learning of Deep Networks from Decentralized Data , 2016, AISTATS.