Expressive Explanations of DNNs by Combining Concept Analysis with ILP
暂无分享,去创建一个
[1] Chun-Liang Li,et al. On Completeness-aware Concept-Based Explanations in Deep Neural Networks , 2020, NeurIPS.
[2] Massimo Mauro,et al. Multi-class semantic segmentation of faces , 2015, 2015 IEEE International Conference on Image Processing (ICIP).
[3] Artur S. d'Avila Garcez,et al. Logic Tensor Networks for Semantic Image Interpretation , 2017, IJCAI.
[4] Geoffrey E. Hinton,et al. Dynamic Routing Between Capsules , 2017, NIPS.
[5] Bettina Finzel,et al. Mutual Explanations for Cooperative Decision Making in Medicine , 2020, KI - Künstliche Intelligenz.
[6] Zhi-Hua Zhou,et al. Bridging Machine Learning and Logical Reasoning by Abductive Learning , 2019, NeurIPS.
[7] Martin Schels,et al. Concept Enforcement and Modularization as Methods for the ISO 26262 Safety Argumentation of Neural Networks , 2020 .
[8] Amina Adadi,et al. Peeking Inside the Black-Box: A Survey on Explainable Artificial Intelligence (XAI) , 2018, IEEE Access.
[9] Alex Krizhevsky,et al. One weird trick for parallelizing convolutional neural networks , 2014, ArXiv.
[10] Andrew Zisserman,et al. Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.
[11] Geoffrey Zweig,et al. Linguistic Regularities in Continuous Space Word Representations , 2013, NAACL.
[12] Michael Siebers,et al. Enriching Visual with Verbal Explanations for Relational Concepts - Combining LIME with Aleph , 2019, PKDD/ECML Workshops.
[13] Ahmad Khatoun,et al. tACS motor system effects can be caused by transcutaneous stimulation of peripheral nerves , 2019, Nature Communications.
[14] Martin Wattenberg,et al. Interpretability Beyond Feature Attribution: Quantitative Testing with Concept Activation Vectors (TCAV) , 2017, ICML.
[15] Massimo Mauro,et al. Head pose estimation through multi-class face segmentation , 2017, 2017 IEEE International Conference on Multimedia and Expo (ICME).
[16] Carlos Guestrin,et al. "Why Should I Trust You?": Explaining the Predictions of Any Classifier , 2016, ArXiv.
[17] A. Antunes. Democracia e Cidadania na Escola: Do Discurso à Prática , 2008 .
[18] Tom M. Mitchell,et al. Explanation-Based Generalization: A Unifying View , 1986, Machine Learning.
[19] Stephen Muggleton,et al. Ultra-Strong Machine Learning: comprehensibility of programs learned with ILP , 2018, Machine Learning.
[20] Amit Dhurandhar,et al. One Explanation Does Not Fit All: A Toolkit and Taxonomy of AI Explainability Techniques , 2019, ArXiv.
[21] Michael Siebers,et al. Explaining Black-Box Classifiers with ILP - Empowering LIME with Aleph to Approximate Non-linear Decisions with Relational Rules , 2018, ILP.
[22] Andrea Vedaldi,et al. Net2Vec: Quantifying and Explaining How Concepts are Encoded by Filters in Deep Neural Networks , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[23] Jun Zhao,et al. Knowledge Graph Embedding via Dynamic Mapping Matrix , 2015, ACL.
[24] Zhuowen Tu,et al. Aggregated Residual Transformations for Deep Neural Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[25] Ute Schmid,et al. Inductive Programming as Approach to Comprehensible Machine Learning , 2018, DKB/KIK@KI.
[26] Ute Schmid,et al. Deep-learned faces of pain and emotions: Elucidating the differences of facial expressions with the help of explainable AI methods , 2019, tm - Technisches Messen.
[27] Klaus-Robert Müller,et al. Explainable artificial intelligence , 2017 .
[28] Alexander Binder,et al. Unmasking Clever Hans predictors and assessing what machines really learn , 2019, Nature Communications.
[29] James Zou,et al. Towards Automatic Concept-based Explanations , 2019, NeurIPS.