R\'enyi Fair Inference

Machine learning algorithms have been increasingly deployed in critical automated decision-making systems that directly affect human lives. When these algorithms are only trained to minimize the training/test error, they could suffer from systematic discrimination against individuals based on their sensitive attributes such as gender or race. Recently, there has been a surge in machine learning society to develop algorithms for fair machine learning. In particular, many adversarial learning procedures have been proposed to impose fairness. Unfortunately, these algorithms either can only impose fairness up to first-order dependence between the variables, or they lack computational convergence guarantees. In this paper, we use Renyi correlation as a measure of fairness of machine learning models and develop a general training framework to impose fairness. In particular, we propose a min-max formulation which balances the accuracy and fairness when solved to optimality. For the case of discrete sensitive attributes, we suggest an iterative algorithm with theoretical convergence guarantee for solving the proposed min-max problem. Our algorithm and analysis are then specialized to fair classification and the fair clustering problem under disparate impact doctrine. Finally, the performance of the proposed Renyi fair inference framework is evaluated on Adult and Bank datasets.

[1]  Michael I. Jordan,et al.  What is Local Optimality in Nonconvex-Nonconcave Minimax Optimization? , 2019, ICML.

[2]  Kush R. Varshney,et al.  Optimized Pre-Processing for Discrimination Prevention , 2017, NIPS.

[3]  Brian D. Ziebart,et al.  Fairness for Robust Log Loss Classification , 2019, AAAI.

[4]  Toniann Pitassi,et al.  Fairness through awareness , 2011, ITCS '12.

[5]  Mingrui Liu,et al.  Non-Convex Min-Max Optimization: Provable Algorithms and Applications in Machine Learning , 2018, ArXiv.

[6]  Yair Carmon,et al.  Lower bounds for finding stationary points I , 2017, Mathematical Programming.

[7]  Michael I. Jordan,et al.  Minmax Optimization: Stable Limit Points of Gradient Descent Ascent are Locally Optimal , 2019, ArXiv.

[8]  Stefano Ermon,et al.  Learning Controllable Fair Representations , 2018, AISTATS.

[9]  Benjamin Fish,et al.  A Confidence-Based Approach for Balancing Fairness and Accuracy , 2016, SDM.

[10]  Toniann Pitassi,et al.  Learning Fair Representations , 2013, ICML.

[11]  Brian D. Ziebart,et al.  Fair Logistic Regression: An Adversarial Perspective , 2019, ArXiv.

[12]  Jason D. Lee,et al.  Solving a Class of Non-Convex Min-Max Games Using Iterative First Order Methods , 2019, NeurIPS.

[13]  Krishna P. Gummadi,et al.  Fairness Constraints: Mechanisms for Fair Classification , 2015, AISTATS.

[14]  J. Danskin The Theory of Max-Min and its Application to Weapons Allocation Problems , 1967 .

[15]  Krzysztof Onak,et al.  Scalable Fair Clustering , 2019, ICML.

[16]  Blake Lemoine,et al.  Mitigating Unwanted Biases with Adversarial Learning , 2018, AIES.

[17]  Salvatore Ruggieri,et al.  Using t-closeness anonymity to control for non-discrimination , 2015, Trans. Data Priv..

[18]  Amos J. Storkey,et al.  Censoring Representations with an Adversary , 2015, ICLR.

[19]  Adam Tauman Kalai,et al.  Unleashing Linear Optimizers for Group-Fair Learning and Optimization , 2018, COLT.

[20]  Jason D. Lee,et al.  On the Convergence and Robustness of Training GANs with Regularized Optimal Transport , 2018, NeurIPS.

[21]  Oliver Kosut,et al.  Tunable Measures for Information Leakage and Applications to Privacy-Utility Tradeoffs , 2018, IEEE Transactions on Information Theory.

[22]  Nisheeth K. Vishnoi,et al.  Classification with Fairness Constraints: A Meta-Algorithm with Provable Guarantees , 2018, FAT.

[23]  Jun Sakuma,et al.  Fairness-aware Learning through Regularization Approach , 2011, 2011 IEEE 11th International Conference on Data Mining Workshops.

[24]  P. Cochat,et al.  Et al , 2008, Archives de pediatrie : organe officiel de la Societe francaise de pediatrie.

[25]  Shai Ben-David,et al.  Empirical Risk Minimization under Fairness Constraints , 2018, NeurIPS.

[26]  Christian Sohler,et al.  Fair Coresets and Streaming Algorithms for Fair k-Means Clustering , 2018, ArXiv.

[27]  Seth Neel,et al.  Preventing Fairness Gerrymandering: Auditing and Learning for Subgroup Fairness , 2017, ICML.

[28]  David Tse,et al.  Discrete Rényi Classifiers , 2015, NIPS.

[29]  Seth Neel,et al.  A Convex Framework for Fair Regression , 2017, ArXiv.

[30]  Valero Laparra,et al.  Fair Kernel Learning , 2017, ECML/PKDD.

[31]  Silvio Lattanzi,et al.  Fair Clustering Through Fairlets , 2018, NIPS.

[32]  A. Rényi On measures of dependence , 1959 .

[33]  Edward Raff,et al.  Gradient Reversal against Discrimination: A Fair Neural Network Learning Approach , 2018, 2018 IEEE 5th International Conference on Data Science and Advanced Analytics (DSAA).

[34]  Krishna P. Gummadi,et al.  Fairness Beyond Disparate Treatment & Disparate Impact: Learning Classification without Disparate Mistreatment , 2016, WWW.

[35]  Katrina Ligett,et al.  Penalizing Unfairness in Binary Classification , 2017 .

[36]  D. Bertsekas Control of uncertain systems with a set-membership description of the uncertainty , 1971 .

[37]  Sreeram Kannan,et al.  Minimum HGR correlation principle: From marginals to joint distribution , 2015, 2015 IEEE International Symposium on Information Theory (ISIT).

[38]  Max Welling,et al.  The Variational Fair Autoencoder , 2015, ICLR.

[39]  H. Hirschfeld A Connection between Correlation and Contingency , 1935, Mathematical Proceedings of the Cambridge Philosophical Society.

[40]  Adam Tauman Kalai,et al.  Man is to Computer Programmer as Woman is to Homemaker? Debiasing Word Embeddings , 2016, NIPS.

[41]  H. Gebelein Das statistische Problem der Korrelation als Variations‐ und Eigenwertproblem und sein Zusammenhang mit der Ausgleichsrechnung , 1941 .

[42]  Aditya Krishna Menon,et al.  The cost of fairness in binary classification , 2018, FAT.

[43]  Lu Zhang,et al.  FairGAN: Fairness-aware Generative Adversarial Networks , 2018, 2018 IEEE International Conference on Big Data (Big Data).

[44]  Nathan Srebro,et al.  Learning Non-Discriminatory Predictors , 2017, COLT.

[45]  Melanie Schmidt,et al.  Privacy preserving clustering with constraints , 2018, ICALP.

[46]  Samir Khuller,et al.  On the cost of essentially fair clusterings , 2018, APPROX-RANDOM.

[47]  John Langford,et al.  A Reductions Approach to Fair Classification , 2018, ICML.

[48]  Toon Calders,et al.  Data preprocessing techniques for classification without discrimination , 2011, Knowledge and Information Systems.

[49]  Toniann Pitassi,et al.  Learning Adversarially Fair and Transferable Representations , 2018, ICML.

[50]  Michael Carl Tschantz,et al.  Automated Experiments on Ad Privacy Settings , 2014, Proc. Priv. Enhancing Technol..

[51]  Bernhard Schölkopf,et al.  Measuring Statistical Dependence with Hilbert-Schmidt Norms , 2005, ALT.

[52]  Adam Tauman Kalai,et al.  Decoupled Classifiers for Group-Fair and Efficient Machine Learning , 2017, FAT.

[53]  Tgk Toon Calders,et al.  Classification with no discrimination by preferential sampling , 2010 .

[54]  Toon Calders,et al.  Classifying without discriminating , 2009, 2009 2nd International Conference on Computer, Control and Communication.

[55]  Toon Calders,et al.  Building Classifiers with Independency Constraints , 2009, 2009 IEEE International Conference on Data Mining Workshops.

[56]  Nathan Srebro,et al.  Equality of Opportunity in Supervised Learning , 2016, NIPS.

[57]  Carlos Eduardo Scheidegger,et al.  Certifying and Removing Disparate Impact , 2014, KDD.

[58]  Kush R. Varshney,et al.  Fairness GAN , 2018, IBM J. Res. Dev..

[59]  Bernhard Schölkopf,et al.  Kernel Methods for Measuring Independence , 2005, J. Mach. Learn. Res..

[60]  Latanya Sweeney,et al.  Discrimination in online ad delivery , 2013, CACM.

[61]  H. Witsenhausen ON SEQUENCES OF PAIRS OF DEPENDENT RANDOM VARIABLES , 1975 .