Fanoos: Multi-Resolution, Multi-Strength, Interactive Explanations for Learned Systems

Machine learning becomes increasingly important to tune or even synthesize the behavior of safety-critical components in highly non-trivial environments, where the inability to understand learned components in general, and neural nets in particular, poses serious obstacles to their adoption. Explainability and interpretability methods for learned systems have gained considerable academic attention, but the focus of current approaches on only one aspect of explanation, at a fixed level of abstraction, and limited if any formal guarantees, prevents those explanations from being digestible by the relevant stakeholders (e.g., end users, certification authorities, engineers) with their diverse backgrounds and situation-specific needs. We introduce Fanoos, a flexible framework for combining formal verification techniques, heuristic search, and user interaction to explore explanations at the desired level of granularity and fidelity. We demonstrate the ability of Fanoos to produce and adjust the abstractness of explanations in response to user requests on a learned controller for an inverted double pendulum and on a learned CPU usage model.

[1]  Franco Turini,et al.  A Survey of Methods for Explaining Black Box Models , 2018, ACM Comput. Surv..

[2]  Or Biran,et al.  Explanation and Justification in Machine Learning : A Survey Or , 2017 .

[3]  Wen Wu,et al.  Towards Developing Verifiable Neural Network Controller , 1996 .

[4]  S. Kambhampati,et al.  Plan Explanation Through Search in an Abstract Model Space * , 2018 .

[5]  Avi Rosenfeld,et al.  A Survey of Interpretability and Explainability in Human-Agent Systems , 2018 .

[6]  D. Bates,et al.  Clinical Decision Support Systems , 1999, Health Informatics.

[7]  George D. Magoulas,et al.  Reliable estimation of a neural network's domain of validity through interval analysis based inversion , 2015, 2015 International Joint Conference on Neural Networks (IJCNN).

[8]  Armin Biere,et al.  Bounded model checking , 2003, Adv. Comput..

[9]  Trevor Darrell,et al.  Textual Explanations for Self-Driving Vehicles , 2018, ECCV.

[10]  Gerhard Friedrich,et al.  A Taxonomy for Generating Explanations in Recommender Systems , 2011, AI Mag..

[11]  Subbarao Kambhampati,et al.  Explicability? Legibility? Predictability? Transparency? Privacy? Security? The Emerging Landscape of Interpretable Agent Behavior , 2018, ICAPS.

[12]  Davide Calvaresi,et al.  Explainable Agents and Robots: Results from a Systematic Literature Review , 2019, AAMAS.

[13]  Alan Fern,et al.  Learning Finite State Representations of Recurrent Policy Networks , 2018, ICLR.

[14]  Edmund M. Clarke,et al.  Counterexample-guided abstraction refinement , 2003, 10th International Symposium on Temporal Representation and Reasoning, 2003 and Fourth International Conference on Temporal Logic. Proceedings..

[15]  William Frawley,et al.  Knowledge Discovery in Databases , 1991 .

[16]  Tameru Hailesilassie,et al.  Rule Extraction Algorithm for Deep Neural Networks: A Review , 2016, ArXiv.

[17]  Henry M. Wellman,et al.  Theory of Mind for Learning and Teaching: The Nature and Role of Explanation. , 2004 .

[18]  Sonia Chernova,et al.  Interactive Hierarchical Task Learning from a Single Demonstration , 2015, 2015 10th ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[19]  Tim Miller,et al.  Explainable AI: Beware of Inmates Running the Asylum Or: How I Learnt to Stop Worrying and Love the Social and Behavioural Sciences , 2017, ArXiv.

[20]  Olaf Stursberg,et al.  Verification of Hybrid Systems Based on Counterexample-Guided Abstraction Refinement , 2003, TACAS.

[21]  Stephanie Rosenthal,et al.  An effective personal mobile robot agent through symbiotic human-robot interaction , 2010, AAMAS.

[22]  Javier García,et al.  A comprehensive survey on safe reinforcement learning , 2015, J. Mach. Learn. Res..

[23]  Sang M. Lee,et al.  A Survey of Decision Support System Applications (1971–April 1988) , 1990 .

[24]  Amina Adadi,et al.  Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI) , 2018, IEEE Access.

[25]  Jörg Rech,et al.  Knowledge Discovery in Databases , 2001, Künstliche Intell..

[26]  C. R. Ramakrishnan,et al.  Proceedings of the Theory and practice of software, 14th international conference on Tools and algorithms for the construction and analysis of systems , 2008 .

[27]  Sebastian Thrun,et al.  Extracting Rules from Artifical Neural Networks with Distributed Representations , 1994, NIPS.

[28]  R. B. Kearfott,et al.  Interval Computations: Introduction, Uses, and Resources , 2000 .

[29]  L. Vogt Game Theory And Pragmatics , 2016 .

[30]  D. Mayne,et al.  Design issues in adaptive control , 1988 .

[31]  Tim Miller,et al.  Explanation in Artificial Intelligence: Insights from the Social Sciences , 2017, Artif. Intell..

[32]  Stephanie Rosenthal,et al.  Dynamic generation and refinement of robot verbalization , 2016, 2016 25th IEEE International Symposium on Robot and Human Interactive Communication (RO-MAN).

[33]  Carlos Guestrin,et al.  "Why Should I Trust You?": Explaining the Predictions of Any Classifier , 2016, ArXiv.

[34]  Sellappan Palaniappan,et al.  Clinical Decision Support Using OLAP With Data Mining , 2008 .

[35]  E. Walter,et al.  Guaranteed characterization of stability domains via set inversion , 1994, IEEE Trans. Autom. Control..

[36]  A. Driescher,et al.  Checking Stability of Neural NARX Models: An Interval Approach , 1997 .

[37]  Zachary Chase Lipton The mythos of model interpretability , 2016, ACM Queue.

[38]  Weiming Xiang,et al.  Reachability Analysis and Safety Verification for Neural Network Control Systems , 2018, ArXiv.

[39]  Jiawei Han,et al.  Discovery of Multiple-Level Association Rules from Large Databases , 1995, VLDB.

[40]  J. Tennyson In the wake of chaos. Unpredictable order in dynamical systems , 1995 .

[41]  Stephen Muggleton,et al.  Inductive Logic Programming: Issues, Results and the Challenge of Learning Language in Logic , 1999, Artif. Intell..

[42]  Tomasz Imielinski,et al.  Mining association rules between sets of items in large databases , 1993, SIGMOD Conference.

[43]  Nikolaj Bjørner,et al.  Z3: An Efficient SMT Solver , 2008, TACAS.

[44]  Luís Torgo,et al.  OpenML: networked science in machine learning , 2014, SKDD.

[45]  E. Shortliffe Clinical decision-support systems , 1990 .

[46]  Wu Wen,et al.  Neuralware engineering: develop verifiable ANN-based systems , 1996, Proceedings IEEE International Joint Symposia on Intelligence and Systems.

[47]  Bradley Hayes,et al.  Improving Robot Controller Transparency Through Autonomous Policy Explanation , 2017, 2017 12th ACM/IEEE International Conference on Human-Robot Interaction (HRI.

[48]  James Bailey,et al.  Advances in Knowledge Discovery and Data Mining , 2016, Lecture Notes in Computer Science.

[49]  Ramakrishnan Srikant,et al.  Mining generalized association rules , 1995, Future Gener. Comput. Syst..

[50]  Junfeng Yang,et al.  Formal Security Analysis of Neural Networks using Symbolic Intervals , 2018, USENIX Security Symposium.

[51]  Anca D. Dragan,et al.  Enabling robots to communicate their objectives , 2017, Autonomous Robots.

[52]  R. Kennedy,et al.  Defense Advanced Research Projects Agency (DARPA). Change 1 , 1996 .

[53]  Luca Pulina,et al.  An Abstraction-Refinement Approach to Verification of Artificial Neural Networks , 2010, CAV.

[54]  Wynne Hsu,et al.  Mining association rules with multiple minimum supports , 1999, KDD '99.

[55]  Mykel J. Kochenderfer,et al.  Reluplex: An Efficient SMT Solver for Verifying Deep Neural Networks , 2017, CAV.

[56]  Stephanie Rosenthal,et al.  CoBots: Robust Symbiotic Autonomous Mobile Service Robots , 2015, IJCAI.

[57]  Luciano Floridi,et al.  The Method of Levels of Abstraction , 2008, Minds and Machines.

[58]  Alexis Papadimitriou,et al.  A generalized taxonomy of explanations styles for traditional and social recommender systems , 2012, Data Mining and Knowledge Discovery.

[59]  E. Kim,et al.  A survey of decision support system applications (1995–2001) , 2006, J. Oper. Res. Soc..

[60]  Patrick Cousot,et al.  Abstract interpretation: a unified lattice model for static analysis of programs by construction or approximation of fixpoints , 1977, POPL.

[61]  HippJochen,et al.  Algorithms for association rule mining a general survey and comparison , 2000 .

[62]  Jeffrey Heer,et al.  Interpretation and trust: designing model-driven visualizations for text analysis , 2012, CHI.

[63]  Sean B. Eom,et al.  A survey of decision support system applications (1988–1994) , 1998, J. Oper. Res. Soc..

[64]  Marcello R. Napolitano,et al.  Verifying Stability of Dynamic Soft-Computing Systems , 1997 .

[65]  S. M. Tan,et al.  Double pendulum: An experiment in chaos , 1993 .

[66]  Joachim Diederich,et al.  Survey and critique of techniques for extracting rules from trained artificial neural networks , 1995, Knowl. Based Syst..

[67]  S. Mohsin,et al.  Neural Networks in Medical Imaging Applications: A Survey , 2013 .

[68]  Brian Scassellati,et al.  Autonomously constructing hierarchical task networks for planning and human-robot collaboration , 2016, 2016 IEEE International Conference on Robotics and Automation (ICRA).

[69]  Robert D. Tennent,et al.  The denotational semantics of programming languages , 1976, CACM.

[70]  Alec Radford,et al.  Proximal Policy Optimization Algorithms , 2017, ArXiv.

[71]  Jiawei Han,et al.  Mining Multiple-Level Association Rules in Large Databases , 1999, IEEE Trans. Knowl. Data Eng..