Beyond Distributive Fairness in Algorithmic Decision Making: Feature Selection for Procedurally Fair Learning

With widespread use of machine learning methods in numerous domains involving humans, several studies have raised questions about the potential for unfairness towards certain individuals or groups. A number of recent works have proposed methods to measure and eliminate unfairness from machine learning models. However, most of this work has focused on only one dimension of fair decision making: distributive fairness, i.e., the fairness of the decision outcomes. In this work, we leverage the rich literature on organizational justice and focus on another dimension of fair decision making: procedural fairness, i.e., the fairness of the decision making process. We propose measures for procedural fairness that consider the input features used in the decision process, and evaluate the moral judgments of humans regarding the use of these features. We operationalize these measures on two real world datasets using human surveys on the Amazon Mechanical Turk (AMT) platform, demonstrating that our measures capture important properties of procedurally fair decision making. We provide fast submodular mechanisms to optimize the tradeoff between procedural fairness and prediction accuracy. On our datasets, we observe empirically that procedural fairness may be achieved with little cost to outcome fairness, but that some loss of accuracy is unavoidable.

[1]  Franco Turini,et al.  k-NN as an implementation of situation testing for discrimination discovery and prevention , 2011, KDD.

[2]  Jun Sakuma,et al.  Fairness-Aware Classifier with Prejudice Remover Regularizer , 2012, ECML/PKDD.

[3]  Zheng Wen,et al.  Optimal Greedy Diversity for Recommendation , 2015, IJCAI.

[4]  J. Greenberg,et al.  A Taxonomy of Organizational Justice Theories , 1987 .

[5]  Tgk Toon Calders,et al.  Classification with no discrimination by preferential sampling , 2010 .

[6]  Hui Lin,et al.  How to select a good training-data subset for transcription: submodular active selection for sequences , 2009, INTERSPEECH.

[7]  Arne Trankell Reliability of evidence : Methods for analyzing and assessing witness statements , 1973 .

[8]  Rishabh K. Iyer,et al.  Submodular Optimization with Submodular Cover and Submodular Knapsack Constraints , 2013, NIPS.

[9]  Krishna P. Gummadi,et al.  Fairness Constraints: Mechanisms for Fair Classification , 2015, AISTATS.

[10]  Andrew D. Selbst,et al.  Big Data's Disparate Impact , 2016 .

[11]  Rishabh K. Iyer,et al.  Algorithms for Approximate Minimization of the Difference Between Submodular Functions, with Applications , 2012, UAI.

[12]  Jon M. Kleinberg,et al.  Inherent Trade-Offs in the Fair Determination of Risk Scores , 2016, ITCS.

[13]  Bernhard Schölkopf,et al.  Avoiding Discrimination through Causal Reasoning , 2017, NIPS.

[14]  J O Beahrs,et al.  Volition, deception, and the evolution of justice. , 1991, The Bulletin of the American Academy of Psychiatry and the Law.

[15]  Toniann Pitassi,et al.  Fairness through awareness , 2011, ITCS '12.

[16]  Krishna P. Gummadi,et al.  Message Impartiality in Social Media Discussions , 2016, ICWSM.

[17]  C. McDermott Discrimination , 2009, Inclusive Equality.

[18]  Peter H. Ditto,et al.  Moral Foundations Theory: The Pragmatic Validity of Moral Pluralism , 2012 .

[19]  Alexandros G. Dimakis,et al.  Restricted Strong Convexity Implies Weak Submodularity , 2016, The Annals of Statistics.

[20]  Sonja B. Starr,et al.  Ban the Box, Criminal Records, and Statistical Discrimination: A Field Experiment , 2016 .

[21]  Jure Leskovec,et al.  A computational approach to politeness with application to social factors , 2013, ACL.

[22]  Andreas Krause,et al.  Near-Optimal Sensor Placements in Gaussian Processes: Theory, Efficient Algorithms and Empirical Studies , 2008, J. Mach. Learn. Res..

[23]  Franco Turini,et al.  Discrimination-aware data mining , 2008, KDD.

[24]  M. Yaari,et al.  On dividing justly , 1984 .

[25]  Toniann Pitassi,et al.  Learning Fair Representations , 2013, ICML.

[26]  Matt J. Kusner,et al.  Counterfactual Fairness , 2017, NIPS.

[27]  Christopher T. Lowenkamp,et al.  False Positives, False Negatives, and False Analyses: A Rejoinder to "Machine Bias: There's Software Used across the Country to Predict Future Criminals. and It's Biased against Blacks" , 2016 .

[28]  Michael D. Buhrmester,et al.  Amazon's Mechanical Turk , 2011, Perspectives on psychological science : a journal of the Association for Psychological Science.

[29]  Mark Fuge,et al.  Diverse Weighted Bipartite b-Matching , 2017, IJCAI.

[30]  Gaël Varoquaux,et al.  Scikit-learn: Machine Learning in Python , 2011, J. Mach. Learn. Res..

[31]  Hui Lin,et al.  A Class of Submodular Functions for Document Summarization , 2011, ACL.

[32]  Siddharth Suri,et al.  Conducting behavioral research on Amazon’s Mechanical Turk , 2010, Behavior research methods.

[33]  Justin M. Rao,et al.  Precinct or Prejudice? Understanding Racial Disparities in New York City's Stop-and-Frisk Policy , 2015 .

[34]  Nathan Srebro,et al.  Equality of Opportunity in Supervised Learning , 2016, NIPS.

[35]  Carlos Eduardo Scheidegger,et al.  Certifying and Removing Disparate Impact , 2014, KDD.

[36]  Krishna P. Gummadi,et al.  Fairness Beyond Disparate Treatment & Disparate Impact: Learning Classification without Disparate Mistreatment , 2016, WWW.

[37]  Sven Strauss Being Good A Short Introduction To Ethics , 2016 .