Using FALCES against bias in automated decisions by integrating fairness in dynamic model ensembles

As regularly reported in the media, automated classifications and decisions based on machine learning models can cause unfair treatment of certain groups of a general population. Classically, the machine learning models are designed to make highly accurate decisions in general. When one machine learning model is not sufficient to define the possibly complex boundary between classes, multiple “specialized” models are used within a model ensemble to further boost accuracy. In particular, dynamic model ensembles pick the most accurate model for each query object, by applying the model that performed best on similar data. Given the labeled data on which models are trained, it is not surprising that any bias possibly present in the data will reflect in the classifiers using the models. To mitigate this, recent work has proposed fair model ensembles, that instead of focusing (solely) on accuracy also optimize global fairness, which is quantified using bias metrics. However, such global fairness that globally minimizes bias may exhibit imbalances in different regions of the data, e.g., caused by some local bias maxima leading to local unfairness. In this paper, we propose to bridge the gap between dynamic model ensembles and fair model ensembles and investigate the problem of devising locally fair and accurate dynamic model ensembles, which ultimately optimize for equal opportunity of similar subjects. Our evaluation shows that our approach outperforms the state-of-the-art for different types and degrees of bias present in training data in terms of both local and global fairness, while reaching comparable accuracy.

[1]  Indre Zliobaite,et al.  Measuring discrimination in algorithmic decision making , 2017, Data Mining and Knowledge Discovery.

[2]  Ankur Taly,et al.  Counterfactual Fairness in Text Classification through Robustness , 2018, AIES.

[3]  Toon Calders,et al.  Classifying without discriminating , 2009, 2009 2nd International Conference on Computer, Control and Communication.

[4]  Toon Calders,et al.  Three naive Bayes approaches for discrimination-free classification , 2010, Data Mining and Knowledge Discovery.

[5]  Ajay Rana,et al.  K-means with Three different Distance Metrics , 2013 .

[6]  R. Polikar,et al.  Ensemble based systems in decision making , 2006, IEEE Circuits and Systems Magazine.

[7]  Kevin W. Bowyer,et al.  Combination of multiple classifiers using local accuracy estimates , 1996, Proceedings CVPR IEEE Computer Society Conference on Computer Vision and Pattern Recognition.

[8]  Zhiqiang Shen,et al.  MEAL: Multi-Model Ensemble via Adversarial Learning , 2018, AAAI.

[9]  Adam Tauman Kalai,et al.  Decoupled Classifiers for Group-Fair and Efficient Machine Learning , 2017, FAT.

[10]  Lin Yang,et al.  A New Ensemble Learning Framework for 3D Biomedical Image Segmentation , 2018, AAAI.

[11]  George D. C. Cavalcanti,et al.  Dynamic classifier selection: Recent advances and perspectives , 2018, Inf. Fusion.

[12]  Vandana,et al.  Survey of Nearest Neighbor Techniques , 2010, ArXiv.

[13]  Franco Turini,et al.  Discrimination-aware data mining , 2008, KDD.

[14]  Cordelia Schmid,et al.  Diversity With Cooperation: Ensemble Methods for Few-Shot Classification , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).

[15]  Jon Louis Bentley,et al.  Multidimensional binary search trees used for associative searching , 1975, CACM.

[16]  Suresh Venkatasubramanian,et al.  On the (im)possibility of fairness , 2016, ArXiv.