Enriching Visual with Verbal Explanations for Relational Concepts - Combining LIME with Aleph

With the increasing number of deep learning applications, there is a growing demand for explanations. Visual explanations provide information about which parts of an image are relevant for a classifier's decision. However, highlighting of image parts (e.g., an eye) cannot capture the relevance of a specific feature value for a class (e.g., that the eye is wide open). Furthermore, highlighting cannot convey whether the classification depends on the mere presence of parts or on a specific spatial relation between them. Consequently, we present an approach that is capable of explaining a classifier's decision in terms of logic rules obtained by the Inductive Logic Programming system Aleph. The examples and the background knowledge needed for Aleph are based on the explanation generation method LIME. We demonstrate our approach with images of a blocksworld domain. First, we show that our approach is capable of identifying a single relation as important explanatory construct. Afterwards, we present the more complex relational concept of towers. Finally, we show how the generated relational rules can be explicitly related with the input image, resulting in richer explanations.

[1]  H. Grice Logic and conversation , 1975 .

[2]  R. Mayer,et al.  For whom is a picture worth a thousand words? Extensions of a dual-coding theory of multimedia learning. , 1994 .

[3]  D. Gentner,et al.  PSYCHOLOGICAL SCIENCE Research Article STRUCTURAL ALIGNMENT IN COMPARISON: No Difference Without Similarity , 2022 .

[4]  Luc De Raedt,et al.  Inductive Logic Programming: Theory and Methods , 1994, J. Log. Program..

[5]  Ashwin Srinivasan,et al.  Theories for Mutagenicity: A Study in First-Order and Feature-Based Induction , 1996, Artif. Intell..

[6]  Jerry Alan Fails,et al.  Interactive machine learning , 2003, IUI '03.

[7]  R. Tibshirani,et al.  Least angle regression , 2004, math/0406456.

[8]  R. Tibshirani,et al.  REJOINDER TO "LEAST ANGLE REGRESSION" BY EFRON ET AL. , 2004, math/0406474.

[9]  Stefano Soatto,et al.  Quick Shift and Kernel Methods for Mode Seeking , 2008, ECCV.

[10]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[11]  Nitish Srivastava,et al.  Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..

[12]  Stephen Muggleton,et al.  Meta-interpretive learning of higher-order dyadic datalog: predicate invention revisited , 2013, Machine Learning.

[13]  Geoffrey E. Hinton,et al.  Deep Learning , 2015, Nature.

[14]  Carlos Guestrin,et al.  "Why Should I Trust You?": Explaining the Predictions of Any Classifier , 2016, ArXiv.

[15]  Sergio Gomez Colmenarejo,et al.  Hybrid computing using a neural network with dynamic external memory , 2016, Nature.

[16]  Michael Siebers,et al.  Characterizing facial expressions by grammars of action unit sequences - A first investigation using ABL , 2016, Inf. Sci..

[17]  Andrew McCallum,et al.  RelNet: End-to-end Modeling of Entities & Relations , 2017, AKBC@NIPS.

[18]  Klaus-Robert Müller,et al.  Explainable artificial intelligence , 2017 .

[19]  Seyed Mehran Kazemi,et al.  RelNN: A Deep Neural Model for Relational Learning , 2017, AAAI.

[20]  Michael Siebers,et al.  Please delete that! Why should I? , 2018, KI - Künstliche Intelligenz.

[21]  Amina Adadi,et al.  Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI) , 2018, IEEE Access.

[22]  Stephen Muggleton,et al.  Ultra-Strong Machine Learning: comprehensibility of programs learned with ILP , 2018, Machine Learning.

[23]  Michael Siebers,et al.  Explaining Black-Box Classifiers with ILP - Empowering LIME with Aleph to Approximate Non-linear Decisions with Relational Rules , 2018, ILP.

[24]  Tim Miller,et al.  Explanation in Artificial Intelligence: Insights from the Social Sciences , 2017, Artif. Intell..