Towards a Rigorous Evaluation of Explainability for Multivariate Time Series

Machine learning-based systems are rapidly gaining popularity and in-line with that there has been a huge research surge in the field of explainability to ensure that machine learning models are reliable, fair, and can be held liable for their decision-making process. Explainable Artificial Intelligence (XAI) methods are typically deployed to debug black-box machine learning models but in comparison to tabular, text, and image data, explainability in time series is still relatively unexplored. The aim of this study was to achieve and evaluate model agnostic explainability in a time series forecasting problem. This work focused on proving a solution for a digital consultancy company aiming to find a data-driven approach in order to understand the effect of their sales related activities on the sales deals closed. The solution involved framing the problem as a time series forecasting problem to predict the sales deals and the explainability was achieved using two novel model agnostic explainability techniques, Local explainable modelagnostic explanations (LIME) and Shapley additive explanations (SHAP) which were evaluated using human evaluation of explainability. The results clearly indicate that the explanations produced by LIME and SHAP greatly helped lay humans in understanding the predictions made by the machine learning model. The presented work can easily be extended to any time series forecasting or classification scenario for achieving explainability and evaluating the produced explanations.

[1]  Been Kim,et al.  Towards A Rigorous Science of Interpretable Machine Learning , 2017, 1702.08608.

[2]  Oluwasanmi Koyejo,et al.  Examples are not enough, learn to criticize! Criticism for Interpretability , 2016, NIPS.

[3]  Kary Främling,et al.  Explainable Agents for Less Bias in Human-Agent Decision Making , 2020, EXTRAAMAS@AAMAS.

[4]  Mohit Bansal,et al.  Evaluating Explainable AI: Which Algorithmic Explanations Help Users Predict Model Behavior? , 2020, ACL.

[5]  Dong Nguyen,et al.  Comparing Automatic and Human Evaluation of Local Explanations for Text Classification , 2018, NAACL.

[6]  José Luis Aznarte,et al.  Shapley additive explanations for NO2 forecasting , 2020, Ecol. Informatics.

[7]  Scott Lundberg,et al.  A Unified Approach to Interpreting Model Predictions , 2017, NIPS.

[8]  Hung-yi Lee,et al.  Temporal pattern attention for multivariate time series forecasting , 2018, Machine Learning.

[9]  Carlos Guestrin,et al.  Model-Agnostic Interpretability of Machine Learning , 2016, ArXiv.

[10]  Houda Bakir,et al.  E-commerce Time Series Forecasting using LSTM Neural Network and Support Vector Regression , 2017, BDIOT2017.

[11]  Nuno Constantino Castro,et al.  Time Series Data Mining , 2009, Encyclopedia of Database Systems.

[12]  Heinrich Taube,et al.  Human Evaluation of Interpretability: The Case of AI-Generated Music Knowledge , 2020, ArXiv.

[13]  Daniel A. Keim,et al.  Towards A Rigorous Evaluation Of XAI Methods On Time Series , 2019, 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW).

[14]  Carlos Guestrin,et al.  "Why Should I Trust You?": Explaining the Predictions of Any Classifier , 2016, ArXiv.

[15]  E. Marubini,et al.  Bravais-Pearson and Spearman Correlation Coefficients: Meaning, Test of Hypothesis and Confidence Interval , 2002 .

[16]  Karim El Mokhtari,et al.  Interpreting financial time series with SHAP values , 2019, CASCON.

[17]  Alexander J. Smola,et al.  Support Vector Regression Machines , 1996, NIPS.

[18]  E Marubini,et al.  Bravais-Pearson and Spearman correlation coefficients: meaning, test of hypothesis and confidence interval. , 2002, The International journal of biological markers.

[19]  Brandon M. Greenwell,et al.  Interpretable Machine Learning , 2019, Hands-On Machine Learning with R.

[20]  Kary Främling,et al.  Explainable Artificial Intelligence Based Heat Recycler Fault Detection in Air Handling Unit , 2019, EXTRAAMAS@AAMAS.

[21]  Christine Dearnley,et al.  A reflection on the use of semi-structured interviews. , 2005, Nurse researcher.

[22]  Balázs Hidasi,et al.  ShiftTree: An Interpretable Model-Based Approach for Time Series Classification , 2011, ECML/PKDD.

[23]  Tim Miller,et al.  Explanation in Artificial Intelligence: Insights from the Social Sciences , 2017, Artif. Intell..

[24]  J. M. Bilbao,et al.  Contributions to the Theory of Games , 2005 .