Unified Group Fairness on Federated Learning

Federated learning (FL) has emerged as an important machine learning paradigm where a global model is trained based on the private data from distributed clients. However, most of existing FL algorithms cannot guarantee the performance fairness towards different groups because of data distribution shift over groups. In this paper, we formulate the problem of unified group fairness on FL, where the groups can be formed by clients (including existing clients and newly added clients) and sensitive attribute(s). To solve this problem, we first propose a general fair federated framework. Then we construct a unified group fairness risk from the view of federated uncertainty set with theoretical analyses to guarantee unified group fairness on FL. We also develop an efficient federated optimization algorithm named Federated Mirror Descent Ascent with Momentum Acceleration (FMDAM) with convergence guarantee. We validate the advantages of the FMDA-M algorithm with various kinds of distribution shift settings in experiments, and the results show that FMDA-M algorithm outperforms the existing fair FL algorithms on unified group fairness. Introduction Federated learning (FL) has emerged as an important machine learning paradigm where distributed clients (e.g., a large number of mobile devices or several organizations) collaboratively train a shared global model while keeping private data on clients (McMahan et al. 2017). However, FL may suffer from fairness problem by disproportionately advantaging or disadvantaging the model performance on different subpopulations, which becomes an increasing concern, especially in some high-stakes scenarios such as loan approvals, healthcare, etc (Kairouz et al. 2019). How to develop a fair FL framework is of paramount importance for both academic research and real applications, and has become an important research theme in recent years (Mohri, Sivek, and Suresh 2019; Li et al. 2019; Wang et al. 2021). A straightforward idea is to apply the methods designed for centralized fairness problem to FL setting. Unfortunately, most of them assume a centralized available training dataset, which infringes the data privacy in FL (Barocas and Selbst 2016; Woodworth et al. 2017; Mehrabi et al. 2021). Recently, some works have been proposed to encourage the federated model to have similar performance over c data of different attributes ai data of newly added clients c′i data of existing clients ci (a) Client-level Fairness (b) Attribute-level Fairness (c) Agnostic Distribution Fairness c1

[1]  Kristina Lerman,et al.  A Survey on Bias and Fairness in Machine Learning , 2019, ACM Comput. Surv..

[2]  Jianzhong Qi,et al.  Federated Learning with Fair Averaging , 2021, IJCAI.

[3]  M. Mohri,et al.  Communication-Efficient Agnostic Federated Averaging , 2021, Interspeech.

[4]  Virginia Smith,et al.  Ditto: Fair and Robust Federated Learning Through Personalization , 2020, ICML.

[5]  Hanghang Tong,et al.  Fairness-aware Agnostic Federated Learning , 2020, SDM.

[6]  Peng Cui,et al.  Stable Adversarial Learning under Distributional Shifts , 2020, AAAI.

[7]  Richard Nock,et al.  Advances and Open Problems in Federated Learning , 2019, Found. Trends Mach. Learn..

[8]  Mehrdad Mahdavi,et al.  Distributionally Robust Federated Averaging , 2021, NeurIPS.

[9]  Yishay Mansour,et al.  Beyond Individual and Group Fairness , 2020, ArXiv.

[10]  Yuekai Sun,et al.  Two Simple Ways to Learn Individual Fairness Metrics from Data , 2020, ICML.

[11]  Tatsunori B. Hashimoto,et al.  Distributionally Robust Neural Networks , 2020, ICLR.

[12]  Sanjiv Kumar,et al.  Federated Learning with Only Positive Labels , 2020, ICML.

[13]  Mehrdad Mahdavi,et al.  Adaptive Personalized Federated Learning , 2020, ArXiv.

[14]  Peter Richtárik,et al.  Federated Learning of a Mixture of Global and Local Models , 2020, ArXiv.

[15]  Deze Zeng,et al.  A Learning-Based Incentive Mechanism for Federated Learning , 2020, IEEE Internet of Things Journal.

[16]  Reuben Binns,et al.  On the apparent conflict between individual and group fairness , 2019, FAT*.

[17]  K. Keutzer,et al.  Multi-source Distilling Domain Adaptation , 2019, AAAI.

[18]  Tian Li,et al.  Fair Resource Allocation in Federated Learning , 2019, ICLR.

[19]  Tatsunori B. Hashimoto,et al.  Distributionally Robust Neural Networks for Group Shifts: On the Importance of Regularization for Worst-Case Generalization , 2019, ArXiv.

[20]  Shengli Xie,et al.  Incentive Mechanism for Reliable Federated Learning: A Joint Optimization Approach to Combining Reputation and Contract Theory , 2019, IEEE Internet of Things Journal.

[21]  Percy Liang,et al.  Distributionally Robust Language Modeling , 2019, EMNLP.

[22]  Aaron Roth,et al.  Average Individual Fairness: Algorithms, Generalization and Experiments , 2019, NeurIPS.

[23]  Tianjian Chen,et al.  Federated Machine Learning: Concept and Applications , 2019 .

[24]  Mehryar Mohri,et al.  Agnostic Federated Learning , 2019, ICML.

[25]  Rui Zhang,et al.  A Hybrid Approach to Privacy-Preserving Federated Learning , 2018, Informatik Spektrum.

[26]  Bo Wang,et al.  Moment Matching for Multi-Source Domain Adaptation , 2018, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).

[27]  Maya R. Gupta,et al.  Optimization with Non-Differentiable Constraints with Applications to Fairness, Recall, Churn, and Other Goals , 2018, J. Mach. Learn. Res..

[28]  John C. Duchi,et al.  Learning Models with Uniform Performance via Distributionally Robust Optimization , 2018, ArXiv.

[29]  Yue Zhao,et al.  Federated Learning with Non-IID Data , 2018, ArXiv.

[30]  Krishna P. Gummadi,et al.  Equity of Attention: Amortizing Individual Fairness in Rankings , 2018, SIGIR.

[31]  Liang Lin,et al.  Deep Cocktail Network: Multi-source Unsupervised Domain Adaptation with Category Shift , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[32]  Alexandra Chouldechova,et al.  A case study of algorithm-assisted decision making in child maltreatment hotline screening decisions , 2018, FAT.

[33]  John C. Duchi,et al.  Certifying Some Distributional Robustness with Principled Adversarial Training , 2017, ICLR.

[34]  Adam Tauman Kalai,et al.  Decoupled Classifiers for Group-Fair and Efficient Machine Learning , 2017, FAT.

[35]  Gang Niu,et al.  Does Distributionally Robust Supervised Learning Give Robust Classifiers? , 2016, ICML.

[36]  Peter Ochs,et al.  Local Convergence of the Heavy-Ball Method and iPiano for Non-convex Optimization , 2016, J. Optim. Theory Appl..

[37]  Daniel Kuhn,et al.  Data-driven distributionally robust optimization using the Wasserstein metric: performance guarantees and tractable reformulations , 2015, Mathematical Programming.

[38]  Roland Vollgraf,et al.  Fashion-MNIST: a Novel Image Dataset for Benchmarking Machine Learning Algorithms , 2017, ArXiv.

[39]  Pramod K. Varshney,et al.  Convergence Analysis of Proximal Gradient with Momentum for Nonconvex Optimization , 2017, ICML.

[40]  Nathan Srebro,et al.  Learning Non-Discriminatory Predictors , 2017, COLT.

[41]  Ananda Theertha Suresh,et al.  Distributed Mean Estimation with Limited Communication , 2016, ICML.

[42]  Alexandra Chouldechova,et al.  Fair prediction with disparate impact: A study of bias in recidivism prediction instruments , 2016, Big Data.

[43]  John C. Duchi,et al.  Variance-based Regularization with Convex Objectives , 2016, NIPS.

[44]  Blaise Agüera y Arcas,et al.  Communication-Efficient Learning of Deep Networks from Decentralized Data , 2016, AISTATS.

[45]  Krishna P. Gummadi,et al.  Fairness Constraints: Mechanisms for Fair Classification , 2015, AISTATS.

[46]  Peter Richtárik,et al.  Federated Learning: Strategies for Improving Communication Efficiency , 2016, ArXiv.

[47]  Peter Richtárik,et al.  Federated Optimization: Distributed Machine Learning for On-Device Intelligence , 2016, ArXiv.

[48]  Nathan Srebro,et al.  Equality of Opportunity in Supervised Learning , 2016, NIPS.

[49]  Andrew D. Selbst,et al.  Big Data's Disparate Impact , 2016 .

[50]  John C. Duchi,et al.  Stochastic Gradient Methods for Distributionally Robust Optimization with f-divergences , 2016, NIPS.

[51]  Daniel Kuhn,et al.  Distributionally Robust Convex Optimization , 2014, Oper. Res..

[52]  Toniann Pitassi,et al.  Learning Fair Representations , 2013, ICML.

[53]  Toniann Pitassi,et al.  Fairness through awareness , 2011, ITCS '12.

[54]  Yinyu Ye,et al.  Distributionally Robust Optimization Under Moment Uncertainty with Application to Data-Driven Problems , 2010, Oper. Res..

[55]  J. Rawls,et al.  Justice as Fairness: A Restatement , 2001 .

[56]  Y. Nesterov A method for solving the convex programming problem with convergence rate O(1/k^2) , 1983 .