A Quantitative Evaluation of Global, Rule-Based Explanations of Post-Hoc, Model Agnostic Methods
暂无分享,去创建一个
[1] Franco Turini,et al. A Survey of Methods for Explaining Black Box Models , 2018, ACM Comput. Surv..
[2] Filip Karlo Dosilovic,et al. Explainable artificial intelligence: A survey , 2018, 2018 41st International Convention on Information and Communication Technology, Electronics and Microelectronics (MIPRO).
[3] Martin Wattenberg,et al. Interpretability Beyond Feature Attribution: Quantitative Testing with Concept Activation Vectors (TCAV) , 2017, ICML.
[4] Jude W. Shavlik,et al. in Advances in Neural Information Processing , 1996 .
[5] Luca Longo,et al. Notions of explainability and evaluation approaches for explainable artificial intelligence , 2021, Inf. Fusion.
[6] Trevor Darrell,et al. Grounding Visual Explanations , 2018, ECCV.
[7] Tim Miller,et al. Explanation in Artificial Intelligence: Insights from the Social Sciences , 2017, Artif. Intell..
[8] Jude W. Shavlik,et al. Using Sampling and Queries to Extract Rules from Trained Neural Networks , 1994, ICML.
[9] Luca Longo,et al. A Qualitative Investigation of the Explainability of Defeasible Argumentation and Non-Monotonic Fuzzy Reasoning , 2018, AICS.
[10] Jure Leskovec,et al. Interpretable Decision Sets: A Joint Framework for Description and Prediction , 2016, KDD.
[11] Lars Niklasson,et al. The Truth is In There - Rule Extraction from Opaque Models Using Genetic Programming , 2004, FLAIRS.
[12] Carlos Guestrin,et al. Anchors: High-Precision Model-Agnostic Explanations , 2018, AAAI.
[13] Osbert Bastani,et al. Interpretability via Model Extraction , 2017, ArXiv.
[14] Alexey Ignatiev,et al. Towards Trustable Explainable AI , 2020, IJCAI.
[15] Serge Guillaume,et al. Designing fuzzy inference systems from data: An interpretability-oriented review , 2001, IEEE Trans. Fuzzy Syst..
[16] Donald C. Wunsch,et al. Neural network explanation using inversion , 2007, Neural Networks.
[17] Yoshua Bengio,et al. Understanding intermediate layers using linear classifier probes , 2016, ICLR.
[18] Chunyan Miao,et al. Building More Explainable Artificial Intelligence With Argumentation , 2018, AAAI.
[19] Freddy Lécué,et al. Explainable Artificial Intelligence: Concepts, Applications, Research Challenges and Visions , 2020, CD-MAKE.
[20] Randy Kerber,et al. ChiMerge: Discretization of Numeric Attributes , 1992, AAAI.
[21] Pinki Roy,et al. Rule Extraction from Training Data Using Neural Network , 2017, Int. J. Artif. Intell. Tools.
[22] M. Stone. Cross‐Validatory Choice and Assessment of Statistical Predictions , 1976 .
[23] Franco Turini,et al. GLocalX - From Local to Global Explanations of Black Box AI Models , 2021, Artif. Intell..
[24] Jung Min Lee,et al. An Integrative 3C evaluation framework for Explainable Artificial Intelligence , 2019, AMCIS.
[25] Glenn Fung,et al. Rule extraction from linear support vector machines , 2005, KDD '05.
[26] Alex A. Freitas,et al. Are we really discovering ''interesting'' knowledge from data? , 2006 .
[27] Mike Wu,et al. Beyond Sparsity: Tree Regularization of Deep Models for Interpretability , 2017, AAAI.
[28] Cynthia Rudin,et al. Interpretable classifiers using rules and Bayesian analysis: Building a better stroke prediction model , 2015, ArXiv.
[29] Pradeep Ravikumar,et al. Representer Point Selection for Explaining Deep Neural Networks , 2018, NeurIPS.
[30] Alex Alves Freitas,et al. On rule interestingness measures , 1999, Knowl. Based Syst..
[31] Francisco Herrera,et al. A study of statistical techniques and performance measures for genetics-based machine learning: accuracy and interpretability , 2009, Soft Comput..
[32] Martin Wattenberg,et al. Visualizing Dataflow Graphs of Deep Learning Models in TensorFlow , 2018, IEEE Transactions on Visualization and Computer Graphics.
[33] Guido Bologna,et al. A Comparison Study on Rule Extraction from Neural Network Ensembles, Boosted Shallow Trees, and SVMs , 2018, Appl. Comput. Intell. Soft Comput..
[34] Douglas Kline,et al. Revisiting squared-error and cross-entropy functions for training neural network classifiers , 2005, Neural Computing & Applications.
[35] Jamal Ahmad Dargham,et al. Finding the number of hidden neurons for an MLP neural network using coarse to fine search technique , 2010, 10th International Conference on Information Science, Signal Processing and their Applications (ISSPA 2010).
[36] Yoshua Bengio,et al. Show, Attend and Tell: Neural Image Caption Generation with Visual Attention , 2015, ICML.
[37] R. R. Hocking,et al. Selection of the Best Subset in Regression Analysis , 1967 .
[38] Mohan S. Kankanhalli,et al. Trends and Trajectories for Explainable, Accountable and Intelligible Systems: An HCI Research Agenda , 2018, CHI.
[39] B. Ripley,et al. Pattern Recognition , 1968, Nature.
[40] T. Kathirvalavakumar,et al. Reverse Engineering the Neural Networks for Rule Extraction in Classification Problems , 2011, Neural Processing Letters.
[41] Jinhee Chun,et al. Post-hoc Explanation using a Mimic Rule for Numerical Data , 2021, ICAART.
[42] Xiting Wang,et al. Towards better analysis of machine learning models: A visual analytics perspective , 2017, Vis. Informatics.
[43] Jin Song Dong,et al. Towards Dependable and Explainable Machine Learning Using Automated Reasoning , 2018, ICFEM.
[44] Alexander M. Rush,et al. LSTMVis: A Tool for Visual Analysis of Hidden State Dynamics in Recurrent Neural Networks , 2016, IEEE Transactions on Visualization and Computer Graphics.
[45] Zhi-Hua Zhou,et al. Medical diagnosis with C4.5 rule preceded by artificial neural network ensemble , 2003, IEEE Transactions on Information Technology in Biomedicine.
[46] Ronald J. Patton,et al. Interpretation of Trained Neural Networks by Rule Extraction , 2001, Fuzzy Days.
[47] Franco Turini,et al. Factual and Counterfactual Explanations for Black Box Decision Making , 2019, IEEE Intelligent Systems.
[48] Jure Leskovec,et al. Interpretable & Explorable Approximations of Black Box Models , 2017, ArXiv.
[49] Luc De Raedt,et al. Neural-Symbolic Learning and Reasoning: Contributions and Challenges , 2015, AAAI Spring Symposia.
[50] José M. Alonso,et al. A Bibliometric Analysis of the Explainable Artificial Intelligence Research Field , 2018, IPMU.
[51] Kai-Uwe Kühnberger,et al. Towards integrated neural–symbolic systems for human-level AI: Two research programs helping to bridge the gaps , 2015, BICA 2015.
[52] José Hernández-Orallo,et al. From Ensemble Methods to Comprehensible Models , 2002, Discovery Science.
[53] Luca Longo,et al. Classification of Explainable Artificial Intelligence Methods through Their Output Formats , 2021, Mach. Learn. Knowl. Extr..
[54] Zhi-Hua Zhou,et al. Extracting symbolic rules from trained neural network ensembles , 2003, AI Commun..
[55] Lars Niklasson,et al. Accuracy vs. comprehensibility in data mining models , 2004 .
[56] Luca Longo,et al. Inferential Models of Mental Workload with Defeasible Argumentation and Non-monotonic Fuzzy Reasoning: a Comparative Study , 2018, AI³@AI*IA.
[57] Tommi S. Jaakkola,et al. On the Robustness of Interpretability Methods , 2018, ArXiv.
[58] Amina Adadi,et al. Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI) , 2018, IEEE Access.
[59] Nitesh V. Chawla,et al. SMOTE: Synthetic Minority Over-sampling Technique , 2002, J. Artif. Intell. Res..
[60] Guido Bologna,et al. Characterization of Symbolic Rules Embedded in Deep DIMLP Networks: A Challenge to Transparency of Deep Learning , 2017, J. Artif. Intell. Soft Comput. Res..