Personalized Federated Learning under Mixture of Distributions

The recent trend towards Personalized Federated Learning (PFL) has garnered significant attention as it allows for the training of models that are tailored to each client while maintaining data privacy. However, current PFL techniques primarily focus on modeling the conditional distribution heterogeneity (i.e. concept shift), which can result in suboptimal performance when the distribution of input data across clients diverges (i.e. covariate shift). Additionally, these techniques often lack the ability to adapt to unseen data, further limiting their effectiveness in real-world scenarios. To address these limitations, we propose a novel approach, FedGMM, which utilizes Gaussian mixture models (GMM) to effectively fit the input data distributions across diverse clients. The model parameters are estimated by maximum likelihood estimation utilizing a federated Expectation-Maximization algorithm, which is solved in closed form and does not assume gradient similarity. Furthermore, FedGMM possesses an additional advantage of adapting to new clients with minimal overhead, and it also enables uncertainty quantification. Empirical evaluations on synthetic and benchmark datasets demonstrate the superior performance of our method in both PFL classification and novel sample detection.

[1]  Wei Cheng,et al.  Personalized Federated Learning via Heterogeneous Modular Networks , 2022, 2022 IEEE International Conference on Data Mining (ICDM).

[2]  É. Moulines,et al.  FedPop: A Bayesian Approach for Personalised Federated Learning , 2022, NeurIPS.

[3]  Xinmei Tian,et al.  DisPFL: Towards Communication-Efficient Personalized Federated Learning via Decentralized Sparse Training , 2022, ICML.

[4]  Ekdeep Singh Lubana,et al.  Orchestra: Unsupervised Federated Learning via Globally Consistent Clustering , 2022, ICML.

[5]  Anit Kumar Sahu,et al.  Self-Aware Personalized Federated Learning , 2022, NeurIPS.

[6]  Weiming Zhuang,et al.  Divergence-aware Federated Self-Supervised Learning , 2022, ICLR.

[7]  Zhangyang Wang,et al.  Efficient Split-Mix Federated Learning for On-Demand and In-Situ Customization , 2022, ICLR.

[8]  Minlie Huang,et al.  Acceleration of Federated Learning with Alleviated Forgetting in Local Training , 2022, ICLR.

[9]  A. Vedaldi,et al.  Open-Set Recognition: A Good Closed-Set Classifier is All You Need , 2021, ICLR.

[10]  G. Fanti,et al.  FedChain: Chained Algorithms for Near-optimal Communication Cost in Federated Learning , 2021, ICLR.

[11]  Tae-Hyun Oh,et al.  FedPara: Low-rank Hadamard Product for Communication-Efficient Federated Learning , 2021, ICLR.

[12]  Qiang Yang,et al.  Towards Personalized Federated Learning , 2021, IEEE Transactions on Neural Networks and Learning Systems.

[13]  K. Ramchandran,et al.  An Efficient Framework for Clustered Federated Learning , 2020, IEEE Transactions on Information Theory.

[14]  Sung Ju Hwang,et al.  Factorized-FL: Personalized Federated Learning with Parameter Factorization & Similarity Matching , 2022, NeurIPS.

[15]  J. Bilmes,et al.  Diverse Client Selection for Federated Learning via Submodular Maximization , 2022, ICLR.

[16]  Giovanni Neglia,et al.  Federated Multi-Task Learning under a Mixture of Distributions , 2021, NeurIPS.

[17]  Shuihai Hu,et al.  HAFLO: GPU-Based Acceleration for Federated Logistic Regression , 2021, ArXiv.

[18]  Quan Z. Sheng,et al.  Communication Efficiency in Federated Learning: Achievements and Challenges , 2021, ArXiv.

[19]  Gal Chechik,et al.  Personalized Federated Learning with Gaussian Processes , 2021, NeurIPS.

[20]  Gu-Yeon Wei,et al.  Gradient Disaggregation: Breaking Privacy in Federated Learning by Reconstructing the User Participant Matrix , 2021, ICML.

[21]  Nuno Vasconcelos,et al.  Learning Deep Classifiers Consistent with Fine-Grained Novelty Detection , 2021, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[22]  Jiayu Zhou,et al.  Data-Free Knowledge Distillation for Heterogeneous Federated Learning , 2021, ICML.

[23]  Shenghui Song,et al.  Hierarchical Quantized Federated Learning: Convergence Analysis and System Design , 2021, ArXiv.

[24]  Ethan Fetaya,et al.  Personalized Federated Learning using Hypernetworks , 2021, ICML.

[25]  Ali Dehghantanha,et al.  A survey on security and privacy of federated learning , 2021, Future Gener. Comput. Syst..

[26]  Karishma Sharma,et al.  Identifying Coordinated Accounts on Social Media through Hidden Influence and Group Behaviours , 2020, KDD.

[27]  Manzil Zaheer,et al.  Adaptive Federated Optimization , 2020, ICLR.

[28]  Richard Nock,et al.  Advances and Open Problems in Federated Learning , 2019, Found. Trends Mach. Learn..

[29]  Wojciech Samek,et al.  Clustered Federated Learning: Model-Agnostic Distributed Multitask Optimization Under Privacy Constraints , 2019, IEEE Transactions on Neural Networks and Learning Systems.

[30]  Yixuan Li,et al.  Energy-based Out-of-distribution Detection , 2020, NeurIPS.

[31]  Hong-You Chen,et al.  FedDistill: Making Bayesian Model Ensemble Applicable to Federated Learning , 2020, ArXiv.

[32]  Reza M. Parizi,et al.  Federated Learning: A Survey on Enabling Technologies, Protocols, and Applications , 2020, IEEE Access.

[33]  Deniz Gündüz,et al.  Federated Learning With Quantized Global Model Updates , 2020, ArXiv.

[34]  Sebastian U. Stich,et al.  Ensemble Distillation for Robust Model Fusion in Federated Learning , 2020, NeurIPS.

[35]  Yonina C. Eldar,et al.  The Communication-Aware Clustered Federated Learning Problem , 2020, 2020 IEEE International Symposium on Information Theory (ISIT).

[36]  Nguyen H. Tran,et al.  Personalized Federated Learning with Moreau Envelopes , 2020, NeurIPS.

[37]  Mehrdad Mahdavi,et al.  Adaptive Personalized Federated Learning , 2020, ArXiv.

[38]  Aryan Mokhtari,et al.  Personalized Federated Learning: A Meta-Learning Approach , 2020, ArXiv.

[39]  Yasaman Khazaeni,et al.  Federated Learning with Matched Averaging , 2020, ICLR.

[40]  Wei Yang Bryan Lim,et al.  Federated Learning in Mobile Edge Networks: A Comprehensive Survey , 2019, IEEE Communications Surveys & Tutorials.

[41]  Anit Kumar Sahu,et al.  Federated Learning: Challenges, Methods, and Future Directions , 2019, IEEE Signal Processing Magazine.

[42]  Klaus-Robert Müller,et al.  Robust and Communication-Efficient Federated Learning From Non-i.i.d. Data , 2019, IEEE Transactions on Neural Networks and Learning Systems.

[43]  Natalia Gimelshein,et al.  PyTorch: An Imperative Style, High-Performance Deep Learning Library , 2019, NeurIPS.

[44]  Hubert Eichner,et al.  Federated Evaluation of On-device Personalization , 2019, ArXiv.

[45]  Jakub Konecný,et al.  Improving Federated Learning Personalization via Model Agnostic Meta Learning , 2019, ArXiv.

[46]  Khe Chai Sim,et al.  An Investigation Into On-device Personalization of End-to-end Automatic Speech Recognition Models , 2019, INTERSPEECH.

[47]  Satya Narayan Shukla,et al.  Interpolation-Prediction Networks for Irregularly Sampled Time Series , 2019, ICLR.

[48]  Taghi M. Khoshgoftaar,et al.  A survey on Image Data Augmentation for Deep Learning , 2019, Journal of Big Data.

[49]  Joachim M. Buhmann,et al.  Variational Federated Multi-Task Learning , 2019, ArXiv.

[50]  Sebastian Caldas,et al.  LEAF: A Benchmark for Federated Settings , 2018, ArXiv.

[51]  Ameet S. Talwalkar,et al.  Federated Kernelized Multi-Task Learning , 2018 .

[52]  Yan Liu,et al.  Variational Recurrent Adversarial Deep Domain Adaptation , 2016, ICLR.

[53]  Marc Tommasi,et al.  Decentralized Collaborative Learning of Personalized Models over Networks , 2016, AISTATS.

[54]  Blaise Agüera y Arcas,et al.  Communication-Efficient Learning of Deep Networks from Decentralized Data , 2016, AISTATS.

[55]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[56]  Alex Krizhevsky,et al.  Learning Multiple Layers of Features from Tiny Images , 2009 .

[57]  Heng Tao Shen,et al.  Principal Component Analysis , 2009, Encyclopedia of Biometrics.