Impartial Predictive Modeling: Ensuring Fairness in Arbitrary Models

Fairness aware data mining aims to prevent algorithms from discriminating against protected groups. The literature has come to an impasse as to what constitutes explainable variability as opposed to discrimination. This stems from incomplete discussions of fairness in statistics. We demonstrate that fairness is achieved by ensuring impartiality with respect to sensitive characteristics. As these characteristics are determined outside of the model, the correct description of the statistical task is to ensure impartiality. We provide a framework for impartiality by accounting for different perspectives on the data generating process. This framework yields a set of impartial estimates that are applicable in a wide variety of situations and post-processing tools to correct estimates from arbitrary models. This effectively separates prediction and fairness goals, allowing modelers to focus on generating highly predictive models without incorporating the constraint of fairness.

[1]  Thomas C. Schelling,et al.  Dynamic models of segregation , 1971 .

[2]  Jun Sakuma,et al.  Fairness-Aware Classifier with Prejudice Remover Regularizer , 2012, ECML/PKDD.

[3]  Aaron Roth,et al.  Fairness in Learning: Classic and Contextual Bandits , 2016, NIPS.

[4]  K. Arrow The Theory of Discrimination , 1971 .

[5]  Dean P. Foster,et al.  Statistics for Business: Decision Making and Analysis , 2010 .

[6]  R. R. Banks,et al.  Race-Based Suspect Selection and Colorblind Equal Protection Doctrine and Discourse , 2002 .

[7]  P. Sahlberg Finnish Lessons: What Can the World Learn from Educational Change in Finland? , 2011 .

[8]  Leo Breiman,et al.  Random Forests , 2001, Machine Learning.

[9]  Peter H. Schuck Groups in a Diverse, Dynamic, Competitive, and Liberal Society: Comments on Owen Fiss's "Groups and the Equal Protection Clause" , 2003 .

[10]  Faisal Kamiran,et al.  Quantifying explainable discrimination and removing illegal discrimination in automated decision making , 2012, Knowledge and Information Systems.

[11]  Joel Brockner,et al.  It's so hard to be fair. , 2006, Harvard business review.

[12]  Franco Turini,et al.  Discrimination-aware data mining , 2008, KDD.

[13]  金泰吉 John Rawls의 社會正義論 , 1982 .

[14]  Trevor Hastie,et al.  The Elements of Statistical Learning , 2001 .

[15]  J. Roemer,et al.  Equality of Opportunity , 2013 .

[16]  Tess Hegedus,et al.  Finnish lessons: What can the world learn from educational change in Finland? , 2012 .

[17]  J. Pearl Causality: Models, Reasoning and Inference , 2000 .

[18]  Josep Domingo-Ferrer,et al.  A Methodology for Direct and Indirect Discrimination Prevention in Data Mining , 2013, IEEE Transactions on Knowledge and Data Engineering.

[19]  Adam Swift,et al.  Legitimate Parental Partiality , 2009 .

[20]  Toon Calders,et al.  Three naive Bayes approaches for discrimination-free classification , 2010, Data Mining and Knowledge Discovery.

[21]  Toon Calders,et al.  Controlling Attribute Effect in Linear Regression , 2013, 2013 IEEE 13th International Conference on Data Mining.

[22]  Richard Zeckhauser,et al.  Racial Profiling , 2019 .

[23]  P. Rosenbaum Design of Observational Studies , 2009, Springer Series in Statistics.

[24]  M. Kearns,et al.  Fairness in Criminal Justice Risk Assessments: The State of the Art , 2017, Sociological Methods & Research.

[25]  P. Holland CAUSATION AND RACE , 2003 .

[26]  Leo Breiman,et al.  Stacked regressions , 2004, Machine Learning.

[27]  Dean P. Foster,et al.  An Economic Argument for Affirmative Action , 1992 .

[28]  E. Phelps The Statistical Theory of Racism and Sexism , 1972 .

[29]  Toniann Pitassi,et al.  Fairness through awareness , 2011, ITCS '12.

[30]  Christopher Winship,et al.  Counterfactuals and Causal Inference: Methods and Principles for Social Research , 2007 .

[31]  Jack M. Balkin,et al.  The American Civil Rights Tradition: Anticlassification or Antisubordination , 2003 .

[32]  Stephen Coate,et al.  Will Affirmative-Action Policies Eliminate Negative Stereotypes? , 1993 .

[33]  Andreas Buja,et al.  Models as Approximations: How Random Predictors and Model Violations Invalidate Classical Inference in Regression , 2014 .

[34]  J. Rawls,et al.  Justice as Fairness: A Restatement , 2001 .

[35]  Toniann Pitassi,et al.  Learning Fair Representations , 2013, ICML.

[36]  A. Goldberger Reverse Regression and Salary Discrimination , 1984 .

[37]  Dean P. Foster,et al.  Submodularity in Statistics: Comparing the Success of Model Selection Methods , 2015, 1510.06301.

[38]  Toon Calders,et al.  Discrimination Aware Decision Tree Learning , 2010, 2010 IEEE International Conference on Data Mining.

[39]  P. Kleingeld,et al.  The Stanford Encyclopedia of Philosophy , 2013 .

[40]  P. H. Beves EQUALITY , 1977, The Lancet.

[41]  Andrew Mason,et al.  Equality of Opportunity, Old and New* , 2001, Ethics.