Identifying and Correcting Label Bias in Machine Learning

Datasets often contain biases which unfairly disadvantage certain groups, and classifiers trained on such datasets can inherit these biases. In this paper, we provide a mathematical formulation of how this bias can arise. We do so by assuming the existence of underlying, unknown, and unbiased labels which are overwritten by an agent who intends to provide accurate labels but may have biases against certain groups. Despite the fact that we only observe the biased labels, we are able to show that the bias may nevertheless be corrected by re-weighting the data points without changing the labels. We show, with theoretical guarantees, that training on the re-weighted dataset corresponds to training on the unobserved but unbiased labels, thus leading to an unbiased machine learning classifier. Our procedure is fast and robust and can be used with virtually any learning algorithm. We evaluate on a number of standard machine learning fairness datasets and a variety of fairness notions, finding that our method outperforms standard approaches in achieving fair classification.

[1]  Michael P. Friedlander,et al.  On minimizing distortion and relative entropy , 2006, IEEE Transactions on Information Theory.

[2]  Stephen Smale,et al.  Finding the Homology of Submanifolds with High Confidence from Random Samples , 2008, Discret. Comput. Geom..

[3]  Franco Turini,et al.  Discrimination-aware data mining , 2008, KDD.

[4]  Toon Calders,et al.  Building Classifiers with Independency Constraints , 2009, 2009 IEEE International Conference on Data Mining Workshops.

[5]  Toon Calders,et al.  Classifying without discriminating , 2009, 2009 2nd International Conference on Computer, Control and Communication.

[6]  Gaël Varoquaux,et al.  Scikit-learn: Machine Learning in Python , 2011, J. Mach. Learn. Res..

[7]  Toon Calders,et al.  Handling Conditional Discrimination , 2011, 2011 IEEE 11th International Conference on Data Mining.

[8]  Toon Calders,et al.  Data preprocessing techniques for classification without discrimination , 2011, Knowledge and Information Systems.

[9]  Dirk P. Kroese,et al.  The Generalized Cross Entropy Method, with Applications to Probability Density Estimation , 2011 .

[10]  Toniann Pitassi,et al.  Fairness through awareness , 2011, ITCS '12.

[11]  Richard D. Phillips,et al.  Information Effect of Entry into Credit Ratings Market: The Case of Insurers' Ratings , 2011 .

[12]  Toniann Pitassi,et al.  Learning Fair Representations , 2013, ICML.

[13]  Sivaraman Balakrishnan,et al.  Cluster Trees on Manifolds , 2013, NIPS.

[14]  Carlos Eduardo Scheidegger,et al.  Certifying and Removing Disparate Impact , 2014, KDD.

[15]  Michael Feldman,et al.  Computational Fairness: Preventing Machine-Learned Discrimination , 2015 .

[16]  Benjamin Fish,et al.  Fair Boosting : a Case Study , 2015 .

[17]  Maya R. Gupta,et al.  Satisfying Real-world Goals with Dataset Constraints , 2016, NIPS.

[18]  Andrew D. Selbst,et al.  Big Data's Disparate Impact , 2016 .

[19]  Amos J. Storkey,et al.  Censoring Representations with an Adversary , 2015, ICLR.

[20]  Adam Tauman Kalai,et al.  Man is to Computer Programmer as Woman is to Homemaker? Debiasing Word Embeddings , 2016, NIPS.

[21]  Nathan Srebro,et al.  Equality of Opportunity in Supervised Learning , 2016, NIPS.

[22]  Kush R. Varshney,et al.  Optimized Pre-Processing for Discrimination Prevention , 2017, NIPS.

[23]  Heinrich Jiang,et al.  Density Level Set Estimation on Manifolds with DBSCAN , 2017, ICML.

[24]  Alexandra Chouldechova,et al.  Fair prediction with disparate impact: A study of bias in recidivism prediction instruments , 2016, Big Data.

[25]  Krishna P. Gummadi,et al.  Fairness Constraints: Mechanisms for Fair Classification , 2015, AISTATS.

[26]  Zhe Zhao,et al.  Data Decisions and Theoretical Implications when Adversarially Learning Fair Representations , 2017, ArXiv.

[27]  Elad Eban,et al.  Scalable Learning of Non-Decomposable Objectives , 2016, AISTATS.

[28]  Nathan Srebro,et al.  Learning Non-Discriminatory Predictors , 2017, COLT.

[29]  Algorithmic decision making and the cost of fairness , 2017, 1701.08230.

[30]  Jon M. Kleinberg,et al.  Inherent Trade-Offs in the Fair Determination of Risk Scores , 2016, ITCS.

[31]  Jon M. Kleinberg,et al.  On Fairness and Calibration , 2017, NIPS.

[32]  Shai Ben-David,et al.  Empirical Risk Minimization under Fairness Constraints , 2018, NeurIPS.

[33]  Akiko Takeda,et al.  Nonconvex Optimization for Regression with Fairness Constraints , 2018, ICML.

[34]  Yiannis Kompatsiaris,et al.  Adaptive Sensitive Reweighting to Mitigate Bias in Fairness-aware Classification , 2018, WWW.

[35]  Seth Neel,et al.  Preventing Fairness Gerrymandering: Auditing and Learning for Subgroup Fairness , 2017, ICML.

[36]  John Langford,et al.  A Reductions Approach to Fair Classification , 2018, ICML.

[37]  Harikrishna Narasimhan,et al.  Learning with Complex Loss Functions and Constraints , 2018, AISTATS.

[38]  Ghassem Tofighi,et al.  Detecting Zones and Threat on 3D Body for Security in Airports using Deep Machine Learning , 2018, ArXiv.

[39]  Bertrand K. Hassani,et al.  Regulatory learning: How to supervise machine learning models? An application to credit scoring , 2018, The Journal of Finance and Data Science.

[40]  Ghassem Tofighi,et al.  Detecting zones and threat on 3D body in security airports using deep learning machine , 2018 .

[41]  Maya R. Gupta,et al.  Optimization with Non-Differentiable Constraints with Applications to Fairness, Recall, Churn, and Other Goals , 2018, J. Mach. Learn. Res..

[42]  Maya R. Gupta,et al.  Training Well-Generalizing Classifiers for Fairness Metrics and Other Data-Dependent Constraints , 2018, ICML.

[43]  Karthik Sridharan,et al.  Two-Player Games for Efficient Non-Convex Constrained Optimization , 2018, ALT.