A Quantitative Evaluation of Global, Rule-Based Explanations of Post-Hoc, Model Agnostic Methods

Understanding the inferences of data-driven, machine-learned models can be seen as a process that discloses the relationships between their input and output. These relationships consist and can be represented as a set of inference rules. However, the models usually do not explicit these rules to their end-users who, subsequently, perceive them as black-boxes and might not trust their predictions. Therefore, scholars have proposed several methods for extracting rules from data-driven machine-learned models to explain their logic. However, limited work exists on the evaluation and comparison of these methods. This study proposes a novel comparative approach to evaluate and compare the rulesets produced by five model-agnostic, post-hoc rule extractors by employing eight quantitative metrics. Eventually, the Friedman test was employed to check whether a method consistently performed better than the others, in terms of the selected metrics, and could be considered superior. Findings demonstrate that these metrics do not provide sufficient evidence to identify superior methods over the others. However, when used together, these metrics form a tool, applicable to every rule-extraction method and machine-learned models, that is, suitable to highlight the strengths and weaknesses of the rule-extractors in various applications in an objective and straightforward manner, without any human interventions. Thus, they are capable of successfully modelling distinctively aspects of explainability, providing to researchers and practitioners vital insights on what a model has learned during its training process and how it makes its predictions.

[1]  Franco Turini,et al.  A Survey of Methods for Explaining Black Box Models , 2018, ACM Comput. Surv..

[2]  Filip Karlo Dosilovic,et al.  Explainable artificial intelligence: A survey , 2018, 2018 41st International Convention on Information and Communication Technology, Electronics and Microelectronics (MIPRO).

[3]  Martin Wattenberg,et al.  Interpretability Beyond Feature Attribution: Quantitative Testing with Concept Activation Vectors (TCAV) , 2017, ICML.

[4]  Jude W. Shavlik,et al.  in Advances in Neural Information Processing , 1996 .

[5]  Luca Longo,et al.  Notions of explainability and evaluation approaches for explainable artificial intelligence , 2021, Inf. Fusion.

[6]  Trevor Darrell,et al.  Grounding Visual Explanations , 2018, ECCV.

[7]  Tim Miller,et al.  Explanation in Artificial Intelligence: Insights from the Social Sciences , 2017, Artif. Intell..

[8]  Jude W. Shavlik,et al.  Using Sampling and Queries to Extract Rules from Trained Neural Networks , 1994, ICML.

[9]  Luca Longo,et al.  A Qualitative Investigation of the Explainability of Defeasible Argumentation and Non-Monotonic Fuzzy Reasoning , 2018, AICS.

[10]  Jure Leskovec,et al.  Interpretable Decision Sets: A Joint Framework for Description and Prediction , 2016, KDD.

[11]  Lars Niklasson,et al.  The Truth is In There - Rule Extraction from Opaque Models Using Genetic Programming , 2004, FLAIRS.

[12]  Carlos Guestrin,et al.  Anchors: High-Precision Model-Agnostic Explanations , 2018, AAAI.

[13]  Osbert Bastani,et al.  Interpretability via Model Extraction , 2017, ArXiv.

[14]  Alexey Ignatiev,et al.  Towards Trustable Explainable AI , 2020, IJCAI.

[15]  Serge Guillaume,et al.  Designing fuzzy inference systems from data: An interpretability-oriented review , 2001, IEEE Trans. Fuzzy Syst..

[16]  Donald C. Wunsch,et al.  Neural network explanation using inversion , 2007, Neural Networks.

[17]  Yoshua Bengio,et al.  Understanding intermediate layers using linear classifier probes , 2016, ICLR.

[18]  Chunyan Miao,et al.  Building More Explainable Artificial Intelligence With Argumentation , 2018, AAAI.

[19]  Freddy Lécué,et al.  Explainable Artificial Intelligence: Concepts, Applications, Research Challenges and Visions , 2020, CD-MAKE.

[20]  Randy Kerber,et al.  ChiMerge: Discretization of Numeric Attributes , 1992, AAAI.

[21]  Pinki Roy,et al.  Rule Extraction from Training Data Using Neural Network , 2017, Int. J. Artif. Intell. Tools.

[22]  M. Stone Cross‐Validatory Choice and Assessment of Statistical Predictions , 1976 .

[23]  Franco Turini,et al.  GLocalX - From Local to Global Explanations of Black Box AI Models , 2021, Artif. Intell..

[24]  Jung Min Lee,et al.  An Integrative 3C evaluation framework for Explainable Artificial Intelligence , 2019, AMCIS.

[25]  Glenn Fung,et al.  Rule extraction from linear support vector machines , 2005, KDD '05.

[26]  Alex A. Freitas,et al.  Are we really discovering ''interesting'' knowledge from data? , 2006 .

[27]  Mike Wu,et al.  Beyond Sparsity: Tree Regularization of Deep Models for Interpretability , 2017, AAAI.

[28]  Cynthia Rudin,et al.  Interpretable classifiers using rules and Bayesian analysis: Building a better stroke prediction model , 2015, ArXiv.

[29]  Pradeep Ravikumar,et al.  Representer Point Selection for Explaining Deep Neural Networks , 2018, NeurIPS.

[30]  Alex Alves Freitas,et al.  On rule interestingness measures , 1999, Knowl. Based Syst..

[31]  Francisco Herrera,et al.  A study of statistical techniques and performance measures for genetics-based machine learning: accuracy and interpretability , 2009, Soft Comput..

[32]  Martin Wattenberg,et al.  Visualizing Dataflow Graphs of Deep Learning Models in TensorFlow , 2018, IEEE Transactions on Visualization and Computer Graphics.

[33]  Guido Bologna,et al.  A Comparison Study on Rule Extraction from Neural Network Ensembles, Boosted Shallow Trees, and SVMs , 2018, Appl. Comput. Intell. Soft Comput..

[34]  Douglas Kline,et al.  Revisiting squared-error and cross-entropy functions for training neural network classifiers , 2005, Neural Computing & Applications.

[35]  Jamal Ahmad Dargham,et al.  Finding the number of hidden neurons for an MLP neural network using coarse to fine search technique , 2010, 10th International Conference on Information Science, Signal Processing and their Applications (ISSPA 2010).

[36]  Yoshua Bengio,et al.  Show, Attend and Tell: Neural Image Caption Generation with Visual Attention , 2015, ICML.

[37]  R. R. Hocking,et al.  Selection of the Best Subset in Regression Analysis , 1967 .

[38]  Mohan S. Kankanhalli,et al.  Trends and Trajectories for Explainable, Accountable and Intelligible Systems: An HCI Research Agenda , 2018, CHI.

[39]  B. Ripley,et al.  Pattern Recognition , 1968, Nature.

[40]  T. Kathirvalavakumar,et al.  Reverse Engineering the Neural Networks for Rule Extraction in Classification Problems , 2011, Neural Processing Letters.

[41]  Jinhee Chun,et al.  Post-hoc Explanation using a Mimic Rule for Numerical Data , 2021, ICAART.

[42]  Xiting Wang,et al.  Towards better analysis of machine learning models: A visual analytics perspective , 2017, Vis. Informatics.

[43]  Jin Song Dong,et al.  Towards Dependable and Explainable Machine Learning Using Automated Reasoning , 2018, ICFEM.

[44]  Alexander M. Rush,et al.  LSTMVis: A Tool for Visual Analysis of Hidden State Dynamics in Recurrent Neural Networks , 2016, IEEE Transactions on Visualization and Computer Graphics.

[45]  Zhi-Hua Zhou,et al.  Medical diagnosis with C4.5 rule preceded by artificial neural network ensemble , 2003, IEEE Transactions on Information Technology in Biomedicine.

[46]  Ronald J. Patton,et al.  Interpretation of Trained Neural Networks by Rule Extraction , 2001, Fuzzy Days.

[47]  Franco Turini,et al.  Factual and Counterfactual Explanations for Black Box Decision Making , 2019, IEEE Intelligent Systems.

[48]  Jure Leskovec,et al.  Interpretable & Explorable Approximations of Black Box Models , 2017, ArXiv.

[49]  Luc De Raedt,et al.  Neural-Symbolic Learning and Reasoning: Contributions and Challenges , 2015, AAAI Spring Symposia.

[50]  José M. Alonso,et al.  A Bibliometric Analysis of the Explainable Artificial Intelligence Research Field , 2018, IPMU.

[51]  Kai-Uwe Kühnberger,et al.  Towards integrated neural–symbolic systems for human-level AI: Two research programs helping to bridge the gaps , 2015, BICA 2015.

[52]  José Hernández-Orallo,et al.  From Ensemble Methods to Comprehensible Models , 2002, Discovery Science.

[53]  Luca Longo,et al.  Classification of Explainable Artificial Intelligence Methods through Their Output Formats , 2021, Mach. Learn. Knowl. Extr..

[54]  Zhi-Hua Zhou,et al.  Extracting symbolic rules from trained neural network ensembles , 2003, AI Commun..

[55]  Lars Niklasson,et al.  Accuracy vs. comprehensibility in data mining models , 2004 .

[56]  Luca Longo,et al.  Inferential Models of Mental Workload with Defeasible Argumentation and Non-monotonic Fuzzy Reasoning: a Comparative Study , 2018, AI³@AI*IA.

[57]  Tommi S. Jaakkola,et al.  On the Robustness of Interpretability Methods , 2018, ArXiv.

[58]  Amina Adadi,et al.  Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI) , 2018, IEEE Access.

[59]  Nitesh V. Chawla,et al.  SMOTE: Synthetic Minority Over-sampling Technique , 2002, J. Artif. Intell. Res..

[60]  Guido Bologna,et al.  Characterization of Symbolic Rules Embedded in Deep DIMLP Networks: A Challenge to Transparency of Deep Learning , 2017, J. Artif. Intell. Soft Comput. Res..