暂无分享,去创建一个
[1] Yoshua Bengio,et al. Generative Adversarial Nets , 2014, NIPS.
[2] Amit Dhurandhar,et al. TIP: Typifying the Interpretability of Procedures , 2017, ArXiv.
[3] Amit Dhurandhar,et al. Generating Contrastive Explanations with Monotonic Attribute Functions , 2019, ArXiv.
[4] Patrick J. F. Groenen,et al. Modern Multidimensional Scaling: Theory and Applications , 2003 .
[5] Alexander Binder,et al. On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation , 2015, PloS one.
[6] Marie-Jeanne Lesot,et al. Comparison-Based Inverse Classification for Interpretability in Machine Learning , 2018, IPMU.
[7] Atsuyoshi Nakamura,et al. Convex sets as prototypes for classifying patterns , 2009, Eng. Appl. Artif. Intell..
[8] Brandon M. Greenwell,et al. Interpretable Machine Learning , 2019, Hands-On Machine Learning with R.
[9] Marc Teboulle,et al. A Fast Iterative Shrinkage-Thresholding Algorithm for Linear Inverse Problems , 2009, SIAM J. Imaging Sci..
[10] Brad Boehmke,et al. Interpretable Machine Learning , 2019 .
[11] Inderjit S. Dhillon,et al. Clustering with Bregman Divergences , 2005, J. Mach. Learn. Res..
[12] Peter J. Rousseeuw,et al. Clustering by means of medoids , 1987 .
[13] H. Zou,et al. Regularization and variable selection via the elastic net , 2005 .
[14] Max Welling,et al. Auto-Encoding Variational Bayes , 2013, ICLR.
[15] Alexander Binder,et al. Explaining nonlinear classification decisions with deep Taylor decomposition , 2015, Pattern Recognit..
[16] Scott Lundberg,et al. A Unified Approach to Interpreting Model Predictions , 2017, NIPS.
[17] Maya R. Gupta,et al. To Trust Or Not To Trust A Classifier , 2018, NeurIPS.
[18] Percy Liang,et al. Understanding Black-box Predictions via Influence Functions , 2017, ICML.
[19] Amit Dhurandhar,et al. Explanations based on the Missing: Towards Contrastive Explanations with Pertinent Negatives , 2018, NeurIPS.
[20] Steven Salzberg,et al. A Weighted Nearest Neighbor Algorithm for Learning with Symbolic Features , 2004, Machine Learning.
[21] Percy Liang,et al. On the Accuracy of Influence Functions for Measuring Group Effects , 2019, NeurIPS.
[22] Amit Sharma,et al. Explaining machine learning classifiers through diverse counterfactual explanations , 2020, FAT*.
[23] R. Tibshirani,et al. Prototype selection for interpretable classification , 2011, 1202.5933.
[24] Carlos Guestrin,et al. "Why Should I Trust You?": Explaining the Predictions of Any Classifier , 2016, ArXiv.
[25] Klaus-Robert Müller,et al. Learning how to explain neural networks: PatternNet and PatternAttribution , 2017, ICLR.
[26] Amit Dhurandhar,et al. Model Agnostic Contrastive Explanations for Structured Data , 2019, ArXiv.
[27] Jon Louis Bentley,et al. Multidimensional binary search trees used for associative searching , 1975, CACM.
[28] Oluwasanmi Koyejo,et al. Interpreting Black Box Predictions using Fisher Kernels , 2018, AISTATS.
[29] Tu Bao Ho,et al. An association-based dissimilarity measure for categorical data , 2005, Pattern Recognit. Lett..
[30] Carlos Guestrin,et al. Anchors: High-Precision Model-Agnostic Explanations , 2018, AAAI.
[31] Bernhard Schölkopf,et al. Kernel Stein Tests for Multiple Model Comparison , 2019, NeurIPS.
[32] Abhishek Kumar,et al. Variational Inference of Disentangled Latent Concepts from Unlabeled Observations , 2017, ICLR.
[33] Oluwasanmi Koyejo,et al. Examples are not enough, learn to criticize! Criticism for Interpretability , 2016, NIPS.
[34] Jaakko Lehtinen,et al. Progressive Growing of GANs for Improved Quality, Stability, and Variation , 2017, ICLR.
[35] Charu C. Aggarwal,et al. Efficient Data Representation by Selecting Prototypes with Importance Weights , 2017, 2019 IEEE International Conference on Data Mining (ICDM).
[36] Karthik S. Gurumoorthy,et al. ProtoDash: Fast Interpretable Prototype Selection , 2017, ArXiv.
[37] Geoffrey E. Hinton,et al. Learning internal representations by error propagation , 1986 .
[38] Martin Wattenberg,et al. Interpretability Beyond Feature Attribution: Quantitative Testing with Concept Activation Vectors (TCAV) , 2017, ICML.
[39] Richard S. Zemel,et al. Prototypical Networks for Few-shot Learning , 2017, NIPS.
[40] Chris Russell,et al. Counterfactual Explanations Without Opening the Black Box: Automated Decisions and the GDPR , 2017, ArXiv.