Fairness-Preserving Empirical Risk Minimization

Author(s): Yang, Guanqun | Advisor(s): Roychowdhury, Vwani P | Abstract: The concerns regarding ramifications of societal bias targeted at a particular identity group (for example, gender or race) residing in algorithmic decision-making systems have been ever-growing in the past decade. It is a common practice of machine learning models' participation in these systems through empirical risk minimization (ERM) principle, which is often the cause of unfairness by trading off underrepresented groups for overall performance. Despite the importance of preserving fairness in such systems, there is hardly consensus in defining unified fairness metrics, designing widely-applicable bias-mitigation algorithms, and delivering interpretable models abiding by the ERM principle. The situation is made even more grievous when non-structural data, including text, image, and audio, is involved in these systems due to the unavailability of the well-defined identity attribute. Current approaches attempt to tackle algorithmic bias in non-structural settings from data itself and intermediate representation together with the inference component within models. In this thesis, we propose to unify all three bias-mitigation operations into one streamlined machine learning pipeline. At the same time, to provide interpretable results, the explorations will be made while carrying out debiasing procedures, and theoretical justifications will be provided accordingly. By ameliorating different bias-mitigation strategies through synergistic effects and addressing model transparency issues by investigating internal representations, we show that the proposed pipeline could provide interpretable machine learning models that embody fairness across different identity groups in numerous non-structural data settings.

[1]  Aaron Roth,et al.  The Ethical Algorithm: The Science of Socially Aware Algorithm Design , 2021, Perspectives on Science and Christian Faith.

[2]  Bias-Resilient Neural Network , 2019, ArXiv.

[3]  Zoubin Ghahramani,et al.  One-Network Adversarial Fairness , 2019, AAAI.

[4]  Andrew McCallum,et al.  Paper Matching with Local Fairness Constraints , 2019, KDD.

[5]  Sahin Cem Geyik,et al.  Fairness-Aware Ranking in Search & Recommendation Systems with Application to LinkedIn Talent Search , 2019, KDD.

[6]  Lucy Vasserman,et al.  Nuanced Metrics for Measuring Unintended Bias with Real Data for Text Classification , 2019, WWW.

[7]  Suresh Venkatasubramanian,et al.  A comparative study of fairness-enhancing interventions in machine learning , 2018, FAT.

[8]  Lucy Vasserman,et al.  Measuring and Mitigating Unintended Bias in Text Classification , 2018, AIES.

[9]  Alexandra Chouldechova,et al.  The Frontiers of Fairness in Machine Learning , 2018, ArXiv.

[10]  Rachel K. E. Bellamy,et al.  AI Fairness 360: An Extensible Toolkit for Detecting, Understanding, and Mitigating Unwanted Algorithmic Bias , 2018, ArXiv.

[11]  John Langford,et al.  A Reductions Approach to Fair Classification , 2018, ICML.

[12]  Timnit Gebru,et al.  Gender Shades: Intersectional Accuracy Disparities in Commercial Gender Classification , 2018, FAT.

[13]  Lianwen Jin,et al.  SCUT-FBP5500: A Diverse Benchmark Dataset for Multi-Paradigm Facial Beauty Prediction , 2018, 2018 24th International Conference on Pattern Recognition (ICPR).

[14]  Tie-Yan Liu,et al.  Adversarial Neural Machine Translation , 2017, ACML.

[15]  M. Kearns,et al.  Fairness in Criminal Justice Risk Assessments: The State of the Art , 2017, Sociological Methods & Research.

[16]  Murray Shanahan,et al.  The Partial Information Decomposition of Generative Neural Network Models , 2017, Entropy.

[17]  Beng Chin Ooi,et al.  Resolving the Bias in Electronic Medical Records , 2017, KDD.

[18]  Jieyu Zhao,et al.  Men Also Like Shopping: Reducing Gender Bias Amplification using Corpus-level Constraints , 2017, EMNLP.

[19]  Zhe Zhao,et al.  Data Decisions and Theoretical Implications when Adversarially Learning Fair Representations , 2017, ArXiv.

[20]  Sebastian Ruder,et al.  An Overview of Multi-Task Learning in Deep Neural Networks , 2017, ArXiv.

[21]  Seth Neel,et al.  A Convex Framework for Fair Regression , 2017, ArXiv.

[22]  Lu Zhang,et al.  Achieving Non-Discrimination in Data Release , 2016, KDD.

[23]  Jon M. Kleinberg,et al.  Inherent Trade-Offs in the Fair Determination of Risk Scores , 2016, ITCS.

[24]  Nathan Srebro,et al.  Equality of Opportunity in Supervised Learning , 2016, NIPS.

[25]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[26]  Sergey Ioffe,et al.  Rethinking the Inception Architecture for Computer Vision , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[27]  Max Welling,et al.  The Variational Fair Autoencoder , 2015, ICLR.

[28]  Toniann Pitassi,et al.  Fairness through awareness , 2011, ITCS '12.

[29]  Koby Crammer,et al.  Analysis of Representations for Domain Adaptation , 2006, NIPS.

[30]  Ron Kohavi,et al.  Scaling Up the Accuracy of Naive-Bayes Classifiers: A Decision-Tree Hybrid , 1996, KDD.