Coping with Mistreatment in Fair Algorithms

Machine learning actively impacts our everyday life in almost all endeavors and domains such as healthcare, finance, and energy. As our dependence on the machine learning increases, it is inevitable that these algorithms will be used to make decisions that will have a direct impact on the society spanning all resolutions from personal choices to world-wide policies. Hence, it is crucial to ensure that (un)intentional bias does not affect the machine learning algorithms especially when they are required to take decisions that may have unintended consequences. Algorithmic fairness techniques have found traction in the machine learning community and many methods and metrics have been proposed to ensure and evaluate fairness in algorithms and data collection. In this paper, we study the algorithmic fairness in a supervised learning setting and examine the effect of optimizing a classifier for the Equal Opportunity metric. We demonstrate that such a classifier has an increased false positive rate across sensitive groups and propose a conceptually simple method to mitigate this bias. We rigorously analyze the proposed method and evaluate it on several real world datasets demonstrating its efficacy. Reproducibility: All source code, and experimental results are available at https://anonymous. 4open.science/r/b6c6653c-5f50-477c-bb13-7dfdeb39d4f4/

[1]  Sara Ahmadian,et al.  Clustering without Over-Representation , 2019, KDD.

[2]  Nathan Srebro,et al.  Equality of Opportunity in Supervised Learning , 2016, NIPS.

[3]  Adam Tauman Kalai,et al.  Decoupled classifiers for fair and efficient machine learning , 2017, ArXiv.

[4]  Prasant Mohapatra,et al.  Fair Algorithms for Hierarchical Agglomerative Clustering , 2020, 2022 21st IEEE International Conference on Machine Learning and Applications (ICMLA).

[5]  G. Wahba,et al.  Some results on Tchebycheffian spline functions , 1971 .

[6]  Ilya Safro,et al.  Fast imbalanced classification of healthcare data with missing values , 2015, 2015 18th International Conference on Information Fusion (Fusion).

[7]  Indre Zliobaite,et al.  On the relation between accuracy and fairness in binary classification , 2015, ArXiv.

[8]  Toniann Pitassi,et al.  Learning Fair Representations , 2013, ICML.

[9]  Krishna P. Gummadi,et al.  Fairness Constraints: Mechanisms for Fair Classification , 2015, AISTATS.

[10]  Benjamin Moseley,et al.  Fair Hierarchical Clustering , 2020, NeurIPS.

[11]  Shai Ben-David,et al.  Empirical Risk Minimization under Fairness Constraints , 2018, NeurIPS.

[12]  Erez Shmueli,et al.  Algorithmic Fairness , 2020, ArXiv.

[13]  Ehsan Sadrfaridpour,et al.  Engineering fast multilevel support vector machines , 2019, Machine Learning.

[14]  Deeparnab Chakrabarty,et al.  Fair Algorithms for Clustering , 2019, NeurIPS.

[15]  Krzysztof Onak,et al.  Scalable Fair Clustering , 2019, ICML.

[16]  Jon M. Kleinberg,et al.  On Fairness and Calibration , 2017, NIPS.

[17]  Toon Calders,et al.  Building Classifiers with Independency Constraints , 2009, 2009 IEEE International Conference on Data Mining Workshops.

[18]  Silvio Lattanzi,et al.  Fair Clustering Through Fairlets , 2018, NIPS.

[19]  Ehsan Sadrfaridpour,et al.  Predictive models for bariatric surgery risks with imbalanced medical datasets , 2019, Ann. Oper. Res..

[20]  Vivek Kothari,et al.  The Final Frontier: Deep Learning in Space , 2020, HotMobile.

[21]  Bingsheng He,et al.  ThunderSVM: A Fast SVM Library on GPUs and CPUs , 2018, J. Mach. Learn. Res..

[22]  Andrew D. Selbst,et al.  Big Data's Disparate Impact , 2016 .

[23]  Krishna P. Gummadi,et al.  Fairness Beyond Disparate Treatment & Disparate Impact: Learning Classification without Disparate Mistreatment , 2016, WWW.

[24]  U. Rajendra Acharya,et al.  Machine Learning in Healthcare Informatics , 2013, Machine Learning in Healthcare Informatics.

[25]  A. Gorban,et al.  The Five Factor Model of personality and evaluation of drug consumption risk , 2015, 1506.06297.