暂无分享,去创建一个
[1] Ivan V. Oseledets,et al. Speeding-up Convolutional Neural Networks Using Fine-tuned CP-Decomposition , 2014, ICLR.
[2] Jian Sun,et al. Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).
[3] Natalia Gimelshein,et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library , 2019, NeurIPS.
[4] Anit Kumar Sahu,et al. Federated Optimization in Heterogeneous Networks , 2018, MLSys.
[5] Xiang Li,et al. On the Convergence of FedAvg on Non-IID Data , 2019, ICLR.
[6] Phillip B. Gibbons,et al. The Non-IID Data Quagmire of Decentralized Machine Learning , 2019, ICML.
[7] Sashank J. Reddi,et al. SCAFFOLD: Stochastic Controlled Averaging for Federated Learning , 2019, ICML.
[8] William J. Dally,et al. Deep Gradient Compression: Reducing the Communication Bandwidth for Distributed Training , 2017, ICLR.
[9] Jie Ding,et al. HeteroFL: Computation and Communication Efficient Federated Learning for Heterogeneous Clients , 2020, ICLR.
[10] Xiaogang Wang,et al. Convolutional neural networks with low-rank regularization , 2015, ICLR.
[11] Sebastian Caldas,et al. LEAF: A Benchmark for Federated Settings , 2018, ArXiv.
[12] Rama Shankar Yadav,et al. A review on energy efficient protocols in wireless sensor networks , 2016, Wirel. Networks.
[13] Sebastian U. Stich,et al. Local SGD Converges Fast and Communicates Little , 2018, ICLR.
[14] Blaise Agüera y Arcas,et al. Communication-Efficient Learning of Deep Networks from Decentralized Data , 2016, AISTATS.
[15] Amos J. Storkey,et al. School of Informatics, University of Edinburgh , 2022 .
[16] Shenghuo Zhu,et al. Parallel Restarted SGD with Faster Convergence and Less Communication: Demystifying Why Model Averaging Works for Deep Learning , 2018, AAAI.
[17] Ramesh Raskar,et al. FedML: A Research Library and Benchmark for Federated Machine Learning , 2020, ArXiv.
[18] Ampalavanapillai Nirmalathas,et al. Modeling the Total Energy Consumption of Mobile Network Services and Applications , 2019, Energies.
[19] Maria Rigaki,et al. A Survey of Privacy Attacks in Machine Learning , 2020, ArXiv.
[20] Sanjiv Kumar,et al. Federated Learning with Only Positive Labels , 2020, ICML.
[21] Peter Richtárik,et al. Federated Learning: Strategies for Improving Communication Efficiency , 2016, ArXiv.
[22] Venkatesh Saligrama,et al. Federated Learning Based on Dynamic Regularization , 2021, ICLR.
[23] Dan Alistarh,et al. The Convergence of Sparsified Gradient Methods , 2018, NeurIPS.
[24] Andrzej Cichocki,et al. Stable Low-rank Tensor Decomposition for Compression of Convolutional Neural Network , 2020, ECCV.
[25] Chong-Min Kyung,et al. Efficient Neural Network Compression , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[26] Cong Xu,et al. TernGrad: Ternary Gradients to Reduce Communication in Distributed Deep Learning , 2017, NIPS.
[27] Aryan Mokhtari,et al. Federated Learning with Compression: Unified Analysis and Sharp Guarantees , 2020, AISTATS.
[28] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[29] Manzil Zaheer,et al. Adaptive Federated Optimization , 2020, ICLR.
[30] Richard Nock,et al. Advances and Open Problems in Federated Learning , 2021, Found. Trends Mach. Learn..
[31] Tengyu Ma,et al. Federated Accelerated Stochastic Gradient Descent , 2020, NeurIPS.
[32] Murali Annavaram,et al. Group Knowledge Transfer: Federated Learning of Large CNNs at the Edge , 2020, NeurIPS.
[33] Jianxin Wu,et al. ThiNet: A Filter Level Pruning Method for Deep Neural Network Compression , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[34] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[35] Sunav Choudhary,et al. Federated Learning with Personalization Layers , 2019, ArXiv.
[36] Aryan Mokhtari,et al. FedPAQ: A Communication-Efficient Federated Learning Method with Periodic Averaging and Quantization , 2019, AISTATS.
[37] Kamyar Azizzadenesheli,et al. signSGD: compressed optimisation for non-convex problems , 2018, ICML.
[38] Yoshua Bengio,et al. Gradient-based learning applied to document recognition , 1998, Proc. IEEE.
[39] Dan Alistarh,et al. QSGD: Communication-Optimal Stochastic Gradient Descent, with Applications to Training Neural Networks , 2016, 1610.02132.
[40] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[41] Andrew Zisserman,et al. Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.
[42] Stephan Günnemann,et al. Introduction to Tensor Decompositions and their Applications in Machine Learning , 2017, ArXiv.