Visual, textual or hybrid: the effect of user expertise on different explanations

As the use of AI algorithms keeps rising continuously, so does the need for their transparency and accountability. However, literature often adopts a one-size-fits-all approach for developing explanations when in practice, the type of explanations needed depends on the type of end-user. This research will look at user expertise as a variable to see how different levels of expertise influence the understanding of explanations. The first iteration consists of developing two common types of explanations (visual and textual explanations) that explain predictions made by a general class of predictive model learners. These explanations are then evaluated by users of different expertise backgrounds to compare the understanding and ease-of-use of each type of explanation with respect to the different expertise groups. Results show strong differences between experts and lay users when using visual and textual explanations, as well as lay users having a preference for visual explanations which they perform significantly worse with. To solve this problem, the second iteration of this research focuses on the shortcomings of the first two explanations and tries to minimize the difference in understanding between both expertise groups. This is done through the means of developing and testing a candidate solution in the form of hybrid explanations, which essentially combine both visual and textual explanations. This hybrid form of explanations shows a significant improvement in terms of correct understanding (for lay users in particular) when compared to visual explanations, whilst not compromising on ease-of-use at the same time.

[1]  Enrico Costanza,et al.  Evaluating saliency map explanations for convolutional neural networks: a user study , 2020, IUI.

[2]  Mireia Ribera,et al.  Can we do better explanations? A proposal of user-centered explainable AI , 2019, IUI Workshops.

[3]  John Riedl,et al.  Explaining collaborative filtering recommendations , 2000, CSCW '00.

[4]  Lalana Kagal,et al.  Explaining Explanations: An Overview of Interpretability of Machine Learning , 2018, 2018 IEEE 5th International Conference on Data Science and Advanced Analytics (DSAA).

[5]  Margaret Burnett Explaining AI: fairly? well? , 2019, IUI Workshops.

[6]  Peter Goos,et al.  Optimal Design of Experiments: A Case Study Approach , 2011 .

[7]  Lee Lacy,et al.  Defense Advanced Research Projects Agency (DARPA) Agent Markup Language Computer Aided Knowledge Acquisition , 2005 .

[8]  Francisco Herrera,et al.  Explainable Artificial Intelligence (XAI): Concepts, Taxonomies, Opportunities and Challenges toward Responsible AI , 2020, Inf. Fusion.

[9]  Alex Endert,et al.  Should We Trust (X)AI? Design Dimensions for Structured Experimental Evaluations , 2020, ArXiv.

[10]  Eric D. Ragan,et al.  A Survey of Evaluation Methods and Measures for Interpretable Machine Learning , 2018, ArXiv.

[11]  Enrico Bertini,et al.  ViCE: visual counterfactual explanations for machine learning models , 2020, IUI.

[12]  Pat Croskerry,et al.  A universal model of diagnostic reasoning. , 2009, Academic medicine : journal of the Association of American Medical Colleges.

[13]  K. Brasel,et al.  Understanding statistical tests. , 2007, Journal of surgical education.

[14]  Peter Brusilovsky,et al.  User modeling and user adapted interaction , 2001 .

[15]  Barbara L. Welther,et al.  The impact , 1995 .

[16]  Dietmar Jannach,et al.  A systematic review and taxonomy of explanations in decision support and recommender systems , 2017, User Modeling and User-Adapted Interaction.

[17]  J. B. Brooke,et al.  SUS: A 'Quick and Dirty' Usability Scale , 1996 .

[18]  Steven M. Drucker,et al.  Gamut: A Design Probe to Understand How Data Scientists Understand Machine Learning Models , 2019, CHI.

[19]  In Choi,et al.  Choosing the Level of Significance: A Decision‐theoretic Approach , 2019, Abacus.

[20]  James T. Miller,et al.  An Empirical Evaluation of the System Usability Scale , 2008, Int. J. Hum. Comput. Interact..

[21]  Judith Masthoff,et al.  A Survey of Explanations in Recommender Systems , 2007, 2007 IEEE 23rd International Conference on Data Engineering Workshop.

[22]  Philip Sedgwick,et al.  Parametric v non-parametric statistical tests , 2012, BMJ : British Medical Journal.

[23]  Qian Yang,et al.  Designing Theory-Driven User-Centric Explainable AI , 2019, CHI.

[24]  Trevor Darrell,et al.  Multimodal Explanations: Justifying Decisions and Pointing to the Evidence , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[25]  Mouzhi Ge,et al.  How should I explain? A comparison of different explanation types for recommender systems , 2014, Int. J. Hum. Comput. Stud..

[26]  Krzysztof Z. Gajos,et al.  Proxy tasks and subjective measures can be misleading in evaluating explainable AI systems , 2020, IUI.

[27]  Markus Neuhäuser,et al.  Wilcoxon Signed Rank Test , 2006 .

[28]  Steven M. Drucker,et al.  TeleGam: Combining Visualization and Verbalization for Interpretable Machine Learning , 2019, 2019 IEEE Visualization Conference (VIS).

[29]  Cuntai Guan,et al.  A Survey on Explainable Artificial Intelligence (XAI): Toward Medical XAI , 2019, IEEE Transactions on Neural Networks and Learning Systems.

[30]  Tim Miller,et al.  Explanation in Artificial Intelligence: Insights from the Social Sciences , 2017, Artif. Intell..

[31]  Béatrice Cahour,et al.  Does projection into use improve trust and exploration? An example with a cruise control system , 2009 .

[32]  Kenney Ng,et al.  Interacting with Predictions: Visual Inspection of Black-box Machine Learning Models , 2016, CHI.

[33]  Geoffrey K Lighthall,et al.  Understanding Decision Making in Critical Care , 2015, Clinical Medicine & Research.

[34]  Peter Brusilovsky,et al.  Beyond the Ranked List: User-Driven Exploration and Diversification of Social Recommendation , 2018, IUI.

[35]  Lise Getoor,et al.  Personalized explanations for hybrid recommender systems , 2019, IUI.

[36]  Amina Adadi,et al.  Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI) , 2018, IEEE Access.