xGAIL: Explainable Generative Adversarial Imitation Learning for Explainable Human Decision Analysis

To make daily decisions, human agents devise their own "strategies" governing their mobility dynamics (e.g., taxi drivers have preferred working regions and times, and urban commuters have preferred routes and transit modes). Recent research such as generative adversarial imitation learning (GAIL) demonstrates successes in learning human decision-making strategies from their behavior data using deep neural networks (DNNs), which can accurately mimic how humans behave in various scenarios, e.g., playing video games, etc. However, such DNN-based models are "black box" models in nature, making it hard to explain what knowledge the models have learned from human, and how the models make such decisions, which was not addressed in the literature of imitation learning. This paper addresses this research gap by proposing xGAIL, the first explainable generative adversarial imitation learning framework. The proposed xGAIL framework consists of two novel components, including Spatial Activation Maximization (SpatialAM) and Spatial Randomized Input Sampling Explanation (SpatialRISE), to extract both global and local knowledge from a well-trained GAIL model that explains how a human agent makes decisions. Especially, we take taxi drivers' passenger-seeking strategy as an example to validate the effectiveness of the proposed xGAIL framework. Our analysis on a large-scale real-world taxi trajectory data shows promising results from two aspects: i) global explainable knowledge of what nearby traffic condition impels a taxi driver to choose a particular direction to find the next passenger, and ii) local explainable knowledge of what key (sometimes hidden) factors a taxi driver considers when making a particular decision.

[1]  T. Kathirvalavakumar,et al.  Reverse Engineering the Neural Networks for Rule Extraction in Classification Problems , 2011, Neural Processing Letters.

[2]  Yu Zheng,et al.  Human-Centric Urban Transit Evaluation and Planning , 2018, 2018 IEEE International Conference on Data Mining (ICDM).

[3]  Stefano Ermon,et al.  Generative Adversarial Imitation Learning , 2016, NIPS.

[4]  Yanhua Li,et al.  Unveiling Taxi Drivers' Strategies via cGAIL: Conditional Generative Adversarial Imitation Learning , 2019, 2019 IEEE International Conference on Data Mining (ICDM).

[5]  Yongjun Chen,et al.  Interpreting Deep Models for Text Analysis via Optimization and Regularization Methods , 2019, AAAI.

[6]  Kate Saenko,et al.  RISE: Randomized Input Sampling for Explanation of Black-box Models , 2018, BMVC.

[7]  Robert Hecht-Nielsen III.3 – Theory of the Backpropagation Neural Network* , 1992 .

[8]  Xia Hu,et al.  Techniques for interpretable machine learning , 2018, Commun. ACM.

[9]  Zhenming Liu,et al.  Dissecting the Learning Curve of Taxi Drivers: A Data-Driven Approach , 2019, SDM.

[10]  A. Getis The Analysis of Spatial Association by Use of Distance Statistics , 2010 .

[11]  Xia Hu,et al.  Representation Interpretation with Spatial Encoding and Multimodal Analytics , 2019, WSDM.

[12]  Chang Yang,et al.  The Rich and the Poor: A Markov Decision Process Approach to Optimizing Taxi Driver Revenue Efficiency , 2016, CIKM.

[13]  Zhenming Liu,et al.  DHPA , 2020, ACM Trans. Intell. Syst. Technol..

[14]  Robert Hecht-Nielsen,et al.  Theory of the backpropagation neural network , 1989, International 1989 Joint Conference on Neural Networks.

[15]  Thomas Brox,et al.  Synthesizing the preferred inputs for neurons in neural networks via deep generator networks , 2016, NIPS.

[16]  Alex X. Liu,et al.  Optimizing Taxi Driver Profit Efficiency: A Spatial Network-Based Markov Decision Process Approach , 2020, IEEE Transactions on Big Data.

[17]  Carlos Guestrin,et al.  Anchors: High-Precision Model-Agnostic Explanations , 2018, AAAI.

[18]  Anind K. Dey,et al.  Modeling Interaction via the Principle of Maximum Causal Entropy , 2010, ICML.

[19]  Keith R. Matthews,et al.  Elementary Linear Algebra , 1998 .

[20]  Carlos Guestrin,et al.  "Why Should I Trust You?": Explaining the Predictions of Any Classifier , 2016, ArXiv.

[21]  Lawrence D. Jackel,et al.  Handwritten Digit Recognition with a Back-Propagation Network , 1989, NIPS.

[22]  Fei-Fei Li,et al.  Visualizing and Understanding Recurrent Networks , 2015, ArXiv.

[23]  Grzegorz Chrupala,et al.  Representation of Linguistic Form and Function in Recurrent Neural Networks , 2016, CL.

[24]  Filip De Turck,et al.  GENESIM: genetic extraction of a single, interpretable model , 2016, NIPS 2016.

[25]  Jan Peters,et al.  Relative Entropy Inverse Reinforcement Learning , 2011, AISTATS.

[26]  Yoshua Bengio,et al.  Generative Adversarial Nets , 2014, NIPS.

[27]  Hod Lipson,et al.  Understanding Neural Networks Through Deep Visualization , 2015, ArXiv.

[28]  Andrew Zisserman,et al.  Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps , 2013, ICLR.

[29]  Sergey Levine,et al.  Learning Robust Rewards with Adversarial Inverse Reinforcement Learning , 2017, ICLR 2017.

[30]  Anind K. Dey,et al.  Maximum Entropy Inverse Reinforcement Learning , 2008, AAAI.

[31]  Eduardo F. Morales,et al.  An Introduction to Reinforcement Learning , 2011 .

[32]  Carlo Ratti,et al.  Revealing Taxi Driver's Mobility Intelligence through His Trace , 2010 .

[33]  Sandjai Bhulai,et al.  Optimization of the revenue of the New York city taxi service using Markov Decision Processes , 2017 .

[34]  Cheng Zeng,et al.  Dynamic Taxi Pricing , 2014, ECAI.