System to Integrate Fairness Transparently: An Industry Approach

There have been significant research efforts to address the issue of unintentional bias in Machine Learning (ML). Many well-known companies have dealt with the fallout after the deployment of their products due to this issue. In an industrial context, enterprises have large-scale ML solutions for a broad class of use cases deployed for different swaths of customers. Trading off the cost of detecting and mitigating bias across this landscape over the lifetime of each use case against the risk of impact to the brand image is a key consideration. We propose a framework for industrial uses that addresses their methodological and mechanization needs. Our approach benefits from prior experience handling security and privacy concerns as well as past internal ML projects. Through significant reuse of bias handling ability at every stage in the ML development lifecycle to guide users we can lower overall costs of reducing bias.

[1]  Inioluwa Deborah Raji,et al.  Closing the AI accountability gap: defining an end-to-end framework for internal algorithmic auditing , 2020, FAT*.

[2]  Nisheeth K. Vishnoi,et al.  Classification with Fairness Constraints: A Meta-Algorithm with Provable Guarantees , 2018, FAT.

[3]  Hanna M. Wallach,et al.  Co-Designing Checklists to Understand Organizational Challenges and Opportunities around Fairness in AI , 2020, CHI.

[4]  Jon M. Kleinberg,et al.  On Fairness and Calibration , 2017, NIPS.

[5]  Krzysztof Onak,et al.  Scalable Fair Clustering , 2019, ICML.

[6]  C. Martin 2015 , 2015, Les 25 ans de l’OMC: Une rétrospective en photos.

[7]  Tom LaGatta,et al.  Conscientious Classification: A Data Scientist's Guide to Discrimination-Aware Classification , 2017, Big Data.

[8]  Andrew D. Selbst,et al.  Big Data's Disparate Impact , 2016 .

[9]  Catherine Tucker,et al.  Algorithmic bias? An empirical study into apparent gender-based discrimination in the display of STEM career ads , 2019 .

[10]  Yuriy Brun,et al.  Fairness testing: testing software for discrimination , 2017, ESEC/SIGSOFT FSE.

[11]  Solon Barocas,et al.  Mitigating Bias in Algorithmic Employment Screening: Evaluating Claims and Practices , 2019, SSRN Electronic Journal.

[12]  Pramodita Sharma 2012 , 2013, Les 25 ans de l’OMC: Une rétrospective en photos.

[13]  Michael Carl Tschantz,et al.  Automated Experiments on Ad Privacy Settings , 2014, Proc. Priv. Enhancing Technol..

[14]  Toniann Pitassi,et al.  Fairness through awareness , 2011, ITCS '12.

[15]  Nanyun Peng,et al.  Debiasing Community Detection: The Importance of Lowly Connected Nodes , 2019, 2019 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining (ASONAM).

[16]  A. James 2010 , 2011, Philo of Alexandria: an Annotated Bibliography 2007-2016.

[17]  Lu Zhang,et al.  A Causal Framework for Discovering and Removing Direct and Indirect Discrimination , 2016, IJCAI.

[18]  Inioluwa Deborah Raji,et al.  Model Cards for Model Reporting , 2018, FAT.

[19]  Michael Veale,et al.  Fairness and Accountability Design Needs for Algorithmic Support in High-Stakes Public Sector Decision-Making , 2018, CHI.

[20]  Timnit Gebru,et al.  Datasheets for datasets , 2018, Commun. ACM.

[21]  Blake Lemoine,et al.  Mitigating Unwanted Biases with Adversarial Learning , 2018, AIES.

[22]  Rachel K. E. Bellamy,et al.  AI Fairness 360: An Extensible Toolkit for Detecting, Understanding, and Mitigating Unwanted Algorithmic Bias , 2018, ArXiv.

[23]  Kush R. Varshney,et al.  Increasing Trust in AI Services through Supplier's Declarations of Conformity , 2018, IBM J. Res. Dev..

[24]  Sriram Vasudevan,et al.  LiFT: A Scalable Framework for Measuring Fairness in ML Applications , 2020, CIKM.

[25]  Kush R. Varshney,et al.  Optimized Pre-Processing for Discrimination Prevention , 2017, NIPS.

[26]  Nisheeth K. Vishnoi,et al.  Toward Controlling Discrimination in Online Ad Auctions , 2019, ICML.

[27]  Toon Calders,et al.  Data preprocessing techniques for classification without discrimination , 2011, Knowledge and Information Systems.

[28]  Julius Adebayo,et al.  FairML : ToolBox for diagnosing bias in predictive modeling , 2016 .

[29]  Michael Carl Tschantz,et al.  Discrimination in Online Advertising: A Multidisciplinary Inquiry , 2018 .

[30]  Nathan Srebro,et al.  Equality of Opportunity in Supervised Learning , 2016, NIPS.

[31]  Carlos Eduardo Scheidegger,et al.  Certifying and Removing Disparate Impact , 2014, KDD.

[32]  Z. Hasan A Survey on Shari’Ah Governance Practices in Malaysia, GCC Countries and the UK , 2011 .

[33]  Jun Sakuma,et al.  Fairness-Aware Classifier with Prejudice Remover Regularizer , 2012, ECML/PKDD.

[34]  Avi Feller,et al.  Algorithmic Decision Making and the Cost of Fairness , 2017, KDD.

[35]  Florence March,et al.  2016 , 2016, Affair of the Heart.

[36]  Roxana Geambasu,et al.  FairTest: Discovering Unwarranted Associations in Data-Driven Applications , 2015, 2017 IEEE European Symposium on Security and Privacy (EuroS&P).

[37]  Krishna P. Gummadi,et al.  Potential for Discrimination in Online Targeted Advertising , 2018, FAT.

[38]  D. Pager,et al.  Meta-analysis of field experiments shows no change in racial discrimination in hiring over time , 2017, Proceedings of the National Academy of Sciences.

[39]  Alexandra Chouldechova,et al.  A snapshot of the frontiers of fairness in machine learning , 2020, Commun. ACM.

[40]  Teva J. Scheer Uniform Guidelines on Employee Selection Procedures , 2007 .

[41]  Kristina Lerman,et al.  A Survey on Bias and Fairness in Machine Learning , 2019, ACM Comput. Surv..

[42]  Xiangliang Zhang,et al.  Decision Theory for Discrimination-Aware Classification , 2012, 2012 IEEE 12th International Conference on Data Mining.

[43]  Toniann Pitassi,et al.  Learning Fair Representations , 2013, ICML.

[44]  Sahin Cem Geyik,et al.  Fairness-Aware Ranking in Search & Recommendation Systems with Application to LinkedIn Talent Search , 2019, KDD.

[45]  Miroslav Dudík,et al.  Improving Fairness in Machine Learning Systems: What Do Industry Practitioners Need? , 2018, CHI.

[46]  Cathy O'Neil,et al.  Conscientious Classification: A Data Scientist's Guide to Discrimination-Aware Classification , 2017, Big Data.

[47]  A. Cavoukian,et al.  Privacy by Design: essential for organizational accountability and strong business practices , 2010 .