Addressing Fairness and Explainability in Image Classification Using Optimal Transport

Algorithmic Fairness and the explainability of potentially unfair outcomes are crucial for establishing trust and accountability of Artificial Intelligence systems in domains such as healthcare and policing. Though significant advances have been made in each of the fields separately, achieving explainability in fairness applications remains challenging, particularly so in domains where deep neural networks are used. At the same time, ethical data-mining has become ever more relevant, as it has been shown countless times that fairness-unaware algorithms result in biased outcomes. Current approaches focus on mitigating biases in the outcomes of the model, but few attempts have been made to try to explain \emph{why} a model is biased. To bridge this gap, we propose a comprehensive approach that leverages optimal transport theory to uncover the causes and implications of biased regions in images, which easily extends to tabular data as well. Through the use of Wasserstein barycenters, we obtain scores that are independent of a sensitive variable but keep their marginal orderings. This step ensures predictive accuracy but also helps us to recover the regions most associated with the generation of the biases. Our findings hold significant implications for the development of trustworthy and unbiased AI systems, fostering transparency, accountability, and fairness in critical decision-making scenarios across diverse domains.

[1]  Arthur Charpentier,et al.  Fairness in Multi-Task Learning via Wasserstein Barycenters , 2023, arXiv.org.

[2]  J. Ser,et al.  Explainable Artificial Intelligence (XAI): What we know and what is left to attain Trustworthy Artificial Intelligence , 2023, Inf. Fusion.

[3]  Evgenii Chzhen,et al.  Fair learning with Wasserstein barycenters for non-decomposable performance measures , 2022, AISTATS.

[4]  D. Zhu,et al.  Counterfactual Interpolation Augmentation (CIA): A Unified Approach to Enhance Fairness and Explainability of DNN , 2022, IJCAI.

[5]  M. Kampel,et al.  Bias and Fairness in Computer Vision Applications of the Criminal Justice System , 2021, 2021 IEEE Symposium Series on Computational Intelligence (SSCI).

[6]  Davide Anguita,et al.  Toward Learning Trustworthily from Data Combining Privacy, Fairness, and Explainability: An Application to Face Recognition , 2021, Entropy.

[7]  Juan E. Gilbert,et al.  Can Explainable AI Explain Unfairness? A Framework for Evaluating Explainable AI , 2021, ArXiv.

[8]  S. Gambs,et al.  Characterizing the risk of fairwashing , 2021, NeurIPS.

[9]  Hyeran Byun,et al.  Learning Disentangled Representation for Fair Facial Attribute Classification via Fairness-aware Information Alignment , 2021, AAAI.

[10]  Aaron Roth,et al.  The Ethical Algorithm: The Science of Socially Aware Algorithm Design , 2021, Perspectives on Science and Christian Faith.

[11]  D. Garreau,et al.  What does LIME really see in images? , 2021, ICML.

[12]  R. Socher,et al.  Deep learning-enabled medical computer vision , 2021, npj Digital Medicine.

[13]  Luca Oneto,et al.  Fair Regression with Wasserstein Barycenters , 2020, NeurIPS.

[14]  Jean-Michel Loubes,et al.  Projection to Fairness in Statistical Learning. , 2020 .

[15]  Kimmo Kärkkäinen,et al.  Gender Slopes: Counterfactual Fairness for Computer Vision Models by Attribute Manipulation , 2020, Proceedings of the 2nd International Workshop on Fairness, Accountability, Transparency and Ethics in Multimedia.

[16]  John Aslanides,et al.  A General Approach to Fairness with Optimal Transport , 2020, AAAI.

[17]  Olga Russakovsky,et al.  Towards Fairness in Visual Recognition: Effective Strategies for Bias Mitigation , 2019, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[18]  Alejandro Barredo Arrieta,et al.  Explainable Artificial Intelligence (XAI): Concepts, Taxonomies, Opportunities and Challenges toward Responsible AI , 2019, Inf. Fusion.

[19]  Fan Yang,et al.  Score-CAM: Score-Weighted Visual Explanations for Convolutional Neural Networks , 2019, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).

[20]  Kristina Lerman,et al.  A Survey on Bias and Fairness in Machine Learning , 2019, ACM Comput. Surv..

[21]  Mukund Sundararajan,et al.  The many Shapley values for model explanation , 2019, ICML.

[22]  Sébastien Gambs,et al.  Fairwashing: the risk of rationalization , 2019, ICML.

[23]  Cynthia Rudin,et al.  Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead , 2018, Nature Machine Intelligence.

[24]  Francisco J. García-Peñalvo,et al.  Proposing a Machine Learning Approach to Analyze and Predict Employment and its Factors , 2018, Int. J. Interact. Multim. Artif. Intell..

[25]  Putra Manggala,et al.  Using Image Fairness Representations in Diversity-Based Re-ranking for Recommendations , 2018, UMAP.

[26]  Tommi S. Jaakkola,et al.  Towards Robust Interpretability with Self-Explaining Neural Networks , 2018, NeurIPS.

[27]  James Y. Zou,et al.  Multiaccuracy: Black-Box Post-Processing for Fairness in Classification , 2018, AIES.

[28]  Timnit Gebru,et al.  Gender Shades: Intersectional Accuracy Disparities in Commercial Gender Classification , 2018, FAT.

[29]  Alexandra Chouldechova,et al.  Does mitigating ML's impact disparity require treatment disparity? , 2017, NeurIPS.

[30]  Anirban Sarkar,et al.  Grad-CAM++: Generalized Gradient-Based Visual Explanations for Deep Convolutional Networks , 2017, 2018 IEEE Winter Conference on Applications of Computer Vision (WACV).

[31]  Scott Lundberg,et al.  A Unified Approach to Interpreting Model Predictions , 2017, NIPS.

[32]  Sven Nyholm,et al.  The Ethics of Accident-Algorithms for Self-Driving Cars: an Applied Trolley Problem? , 2016, Ethical Theory and Moral Practice.

[33]  Abhishek Das,et al.  Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization , 2016, 2017 IEEE International Conference on Computer Vision (ICCV).

[34]  Marco Tulio Ribeiro,et al.  "Why Should I Trust You?": Explaining the Predictions of Any Classifier , 2016, HLT-NAACL Demos.

[35]  Xiaogang Wang,et al.  Deep Learning Face Attributes in the Wild , 2014, 2015 IEEE International Conference on Computer Vision (ICCV).

[36]  Noah J. Goodall,et al.  Ethical Decision Making during Automated Vehicle Crashes , 2014, ArXiv.

[37]  R. Berk Criminal Justice Forecasts of Risk: A Machine Learning Approach , 2012 .

[38]  Toniann Pitassi,et al.  Fairness through awareness , 2011, ITCS '12.

[39]  Toon Calders,et al.  Building Classifiers with Independency Constraints , 2009, 2009 IEEE International Conference on Data Mining Workshops.

[40]  C. Villani Optimal Transport: Old and New , 2008 .

[41]  Franco Turini,et al.  Discrimination-aware data mining , 2008, KDD.

[42]  L. Shapley A Value for n-person Games , 1988 .

[43]  Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence, IJCAI 2022, Vienna, Austria, 23-29 July 2022 , 2022, IJCAI.

[44]  Filippo Santambrogio,et al.  Optimal Transport for Applied Mathematicians , 2015 .