CyclicFL: A Cyclic Model Pre-Training Approach to Efficient Federated Learning
暂无分享,去创建一个
Yingbo Zhou | Xian Wei | Mingsong Chen | Ming Hu | Xin Fu | Peng Zhang
[1] J. Zhang,et al. FedTune: A Deep Dive into Efficient Federated Fine-Tuning with Pre-trained Transformers , 2022, ArXiv.
[2] Dongmei Zhang,et al. Neuron Campaign for Initialization Guided by Information Bottleneck Theory , 2021, CIKM.
[3] Hangyu Zhu,et al. Federated Learning on Non-IID Data: A Survey , 2021, Neurocomputing.
[4] Ilya Sutskever,et al. Learning Transferable Visual Models From Natural Language Supervision , 2021, ICML.
[5] Tom Goldstein,et al. GradInit: Learning to Initialize Neural Networks for Stable and Efficient Training , 2021, NeurIPS.
[6] Bingsheng He,et al. Federated Learning on Non-IID Data Silos: An Experimental Study , 2021, 2022 IEEE 38th International Conference on Data Engineering (ICDE).
[7] Tongquan Wei,et al. Efficient Federated Learning for Cloud-Based AIoT Applications , 2020, IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems.
[8] S. Gelly,et al. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale , 2020, ICLR.
[9] Spyridon Bakas,et al. Federated learning in medicine: facilitating multi-institutional collaborations without sharing patient data , 2020, Scientific Reports.
[10] Qinghua Liu,et al. Tackling the Objective Inconsistency Problem in Heterogeneous Federated Optimization , 2020, NeurIPS.
[11] Masashi Sugiyama,et al. Generalisation Guarantees for Continual Learning with Orthogonal Gradient Descent , 2020, ArXiv.
[12] Jeffrey Pennington,et al. Provable Benefit of Orthogonal Initialization in Optimizing Deep Linear Networks , 2020, ICLR.
[13] Sashank J. Reddi,et al. SCAFFOLD: Stochastic Controlled Averaging for Federated Learning , 2019, ICML.
[14] Michael I. Jordan,et al. Towards Understanding the Transferability of Deep Representations , 2019, ArXiv.
[15] Kejiang Ye,et al. FFD: A Federated Learning Based Method for Credit Card Fraud Detection , 2019, BigData.
[16] Song Han,et al. Deep Leakage from Gradients , 2019, NeurIPS.
[17] Anit Kumar Sahu,et al. Federated Optimization in Heterogeneous Networks , 2018, MLSys.
[18] Sebastian Caldas,et al. LEAF: A Benchmark for Federated Settings , 2018, ArXiv.
[19] Stephen Marshall,et al. Activation Functions: Comparison of trends in Practice and Research for Deep Learning , 2018, ArXiv.
[20] Barnabás Póczos,et al. Gradient Descent Provably Optimizes Over-parameterized Neural Networks , 2018, ICLR.
[21] Arthur Jacot,et al. Neural Tangent Kernel: Convergence and Generalization in Neural Networks , 2018, NeurIPS.
[22] Sebastian Ruder,et al. Universal Language Model Fine-tuning for Text Classification , 2018, ACL.
[23] Hao Li,et al. Visualizing the Loss Landscape of Neural Nets , 2017, NeurIPS.
[24] Bo Chen,et al. MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications , 2017, ArXiv.
[25] Blaise Agüera y Arcas,et al. Communication-Efficient Learning of Deep Networks from Decentralized Data , 2016, AISTATS.
[26] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[27] Geoffrey E. Hinton,et al. Distilling the Knowledge in a Neural Network , 2015, ArXiv.
[28] Jian Sun,et al. Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).
[29] Andrew Zisserman,et al. Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.
[30] Geoffrey E. Hinton,et al. Rectified Linear Units Improve Restricted Boltzmann Machines , 2010, ICML.
[31] Yoshua Bengio,et al. Understanding the difficulty of training deep feedforward neural networks , 2010, AISTATS.
[32] A. Savitzky,et al. Smoothing and Differentiation of Data by Simplified Least Squares Procedures. , 1964 .
[33] Wei-Lun Chao,et al. On Pre-Training for Federated Learning , 2022, ArXiv.
[34] Michael G. Rabbat,et al. Where to Begin? Exploring the Impact of Pre-Training and Initialization in Federated Learning , 2022, ArXiv.
[35] Ming-Wei Chang,et al. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding , 2019, NAACL.
[36] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[37] Yoshua Bengio,et al. Gradient-based learning applied to document recognition , 1998, Proc. IEEE.