Rényi Fair Inference

Machine learning algorithms have been increasingly deployed in critical automated decision-making systems that directly affect human lives. When these algorithms are solely trained to minimize the training/test error, they could suffer from systematic discrimination against individuals based on their sensitive attributes, such as gender or race. Recently, there has been a surge in machine learning society to develop algorithms for fair machine learning. In particular, several adversarial learning procedures have been proposed to impose fairness. Unfortunately, these algorithms either can only impose fairness up to linear dependence between the variables, or they lack computational convergence guarantees. In this paper, we use Renyi correlation as a measure of fairness of machine learning models and develop a general training framework to impose fairness. In particular, we propose a min-max formulation which balances the accuracy and fairness when solved to optimality. For the case of discrete sensitive attributes, we suggest an iterative algorithm with theoretical convergence guarantee for solving the proposed min-max problem. Our algorithm and analysis are then specialized to fair classification and fair clustering problems. To demonstrate the performance of the proposed Renyi fair inference framework in practice, we compare it with well-known existing methods on several benchmark datasets. Experiments indicate that the proposed method has favorable empirical performance against state-of-the-art approaches.

[1]  Lu Zhang,et al.  FairGAN: Fairness-aware Generative Adversarial Networks , 2018, 2018 IEEE International Conference on Big Data (Big Data).

[2]  Nathan Srebro,et al.  Learning Non-Discriminatory Predictors , 2017, COLT.

[3]  Melanie Schmidt,et al.  Privacy preserving clustering with constraints , 2018, ICALP.

[4]  Yair Carmon,et al.  Lower bounds for finding stationary points I , 2017, Mathematical Programming.

[5]  Michael I. Jordan,et al.  Minmax Optimization: Stable Limit Points of Gradient Descent Ascent are Locally Optimal , 2019, ArXiv.

[6]  Michael Carl Tschantz,et al.  Automated Experiments on Ad Privacy Settings , 2014, Proc. Priv. Enhancing Technol..

[7]  Stefano Ermon,et al.  Learning Controllable Fair Representations , 2018, AISTATS.

[8]  H. Hirschfeld A Connection between Correlation and Contingency , 1935, Mathematical Proceedings of the Cambridge Philosophical Society.

[9]  H. Gebelein Das statistische Problem der Korrelation als Variations‐ und Eigenwertproblem und sein Zusammenhang mit der Ausgleichsrechnung , 1941 .

[10]  Toniann Pitassi,et al.  Learning Fair Representations , 2013, ICML.

[11]  Toon Calders,et al.  Building Classifiers with Independency Constraints , 2009, 2009 IEEE International Conference on Data Mining Workshops.

[12]  Krzysztof Onak,et al.  Scalable Fair Clustering , 2019, ICML.

[13]  Silvio Lattanzi,et al.  Fair Clustering Through Fairlets , 2018, NIPS.

[14]  Toniann Pitassi,et al.  Fairness through awareness , 2011, ITCS '12.

[15]  Mingrui Liu,et al.  Non-Convex Min-Max Optimization: Provable Algorithms and Applications in Machine Learning , 2018, ArXiv.

[16]  John Langford,et al.  A Reductions Approach to Fair Classification , 2018, ICML.

[17]  Max Welling,et al.  The Variational Fair Autoencoder , 2015, ICLR.

[18]  Katrina Ligett,et al.  Penalizing Unfairness in Binary Classification , 2017 .

[19]  D. Bertsekas Control of uncertain systems with a set-membership description of the uncertainty , 1971 .

[20]  Amos J. Storkey,et al.  Censoring Representations with an Adversary , 2015, ICLR.

[21]  Adam Tauman Kalai,et al.  Decoupled Classifiers for Group-Fair and Efficient Machine Learning , 2017, FAT.

[22]  Tgk Toon Calders,et al.  Classification with no discrimination by preferential sampling , 2010 .

[23]  Nisheeth K. Vishnoi,et al.  Classification with Fairness Constraints: A Meta-Algorithm with Provable Guarantees , 2018, FAT.

[24]  Jun Sakuma,et al.  Fairness-aware Learning through Regularization Approach , 2011, 2011 IEEE 11th International Conference on Data Mining Workshops.

[25]  David Tse,et al.  Discrete Rényi Classifiers , 2015, NIPS.

[26]  Samir Khuller,et al.  On the cost of essentially fair clusterings , 2018, APPROX-RANDOM.

[27]  Adam Tauman Kalai,et al.  Man is to Computer Programmer as Woman is to Homemaker? Debiasing Word Embeddings , 2016, NIPS.

[28]  Aditya Krishna Menon,et al.  The cost of fairness in binary classification , 2018, FAT.

[29]  H. Witsenhausen ON SEQUENCES OF PAIRS OF DEPENDENT RANDOM VARIABLES , 1975 .

[30]  Benjamin Fish,et al.  A Confidence-Based Approach for Balancing Fairness and Accuracy , 2016, SDM.

[31]  Toon Calders,et al.  Classifying without discriminating , 2009, 2009 2nd International Conference on Computer, Control and Communication.

[32]  Jason D. Lee,et al.  On the Convergence and Robustness of Training GANs with Regularized Optimal Transport , 2018, NeurIPS.

[33]  Oliver Kosut,et al.  Tunable Measures for Information Leakage and Applications to Privacy-Utility Tradeoffs , 2018, IEEE Transactions on Information Theory.

[34]  Seth Neel,et al.  A Convex Framework for Fair Regression , 2017, ArXiv.

[35]  Kush R. Varshney,et al.  Optimized Pre-Processing for Discrimination Prevention , 2017, NIPS.

[36]  Bernhard Schölkopf,et al.  Measuring Statistical Dependence with Hilbert-Schmidt Norms , 2005, ALT.

[37]  Shai Ben-David,et al.  Empirical Risk Minimization under Fairness Constraints , 2018, NeurIPS.

[38]  Christian Sohler,et al.  Fair Coresets and Streaming Algorithms for Fair k-Means Clustering , 2018, ArXiv.

[39]  Seth Neel,et al.  Preventing Fairness Gerrymandering: Auditing and Learning for Subgroup Fairness , 2017, ICML.

[40]  Adam Tauman Kalai,et al.  Unleashing Linear Optimizers for Group-Fair Learning and Optimization , 2018, COLT.

[41]  P. Cochat,et al.  Et al , 2008, Archives de pediatrie : organe officiel de la Societe francaise de pediatrie.

[42]  Brian D. Ziebart,et al.  Fair Logistic Regression: An Adversarial Perspective , 2019, ArXiv.

[43]  Jason D. Lee,et al.  Solving a Class of Non-Convex Min-Max Games Using Iterative First Order Methods , 2019, NeurIPS.

[44]  Krishna P. Gummadi,et al.  Fairness Constraints: Mechanisms for Fair Classification , 2015, AISTATS.

[45]  J. Danskin The Theory of Max-Min and its Application to Weapons Allocation Problems , 1967 .

[46]  Nathan Srebro,et al.  Equality of Opportunity in Supervised Learning , 2016, NIPS.

[47]  Carlos Eduardo Scheidegger,et al.  Certifying and Removing Disparate Impact , 2014, KDD.

[48]  Kush R. Varshney,et al.  Fairness GAN , 2018, IBM J. Res. Dev..

[49]  Valero Laparra,et al.  Fair Kernel Learning , 2017, ECML/PKDD.

[50]  Sreeram Kannan,et al.  Minimum HGR correlation principle: From marginals to joint distribution , 2015, 2015 IEEE International Symposium on Information Theory (ISIT).

[51]  Blake Lemoine,et al.  Mitigating Unwanted Biases with Adversarial Learning , 2018, AIES.

[52]  Bernhard Schölkopf,et al.  Kernel Methods for Measuring Independence , 2005, J. Mach. Learn. Res..

[53]  Latanya Sweeney,et al.  Discrimination in online ad delivery , 2013, CACM.

[54]  Salvatore Ruggieri,et al.  Using t-closeness anonymity to control for non-discrimination , 2015, Trans. Data Priv..

[55]  A. Rényi On measures of dependence , 1959 .

[56]  Edward Raff,et al.  Gradient Reversal against Discrimination: A Fair Neural Network Learning Approach , 2018, 2018 IEEE 5th International Conference on Data Science and Advanced Analytics (DSAA).

[57]  Krishna P. Gummadi,et al.  Fairness Beyond Disparate Treatment & Disparate Impact: Learning Classification without Disparate Mistreatment , 2016, WWW.

[58]  Toon Calders,et al.  Data preprocessing techniques for classification without discrimination , 2011, Knowledge and Information Systems.

[59]  Toniann Pitassi,et al.  Learning Adversarially Fair and Transferable Representations , 2018, ICML.