Explaining with Counter Visual Attributes and Examples
暂无分享,去创建一个
[1] Alexei A. Efros,et al. What makes Paris look like Paris? , 2015, Commun. ACM.
[2] Michael S. Bernstein,et al. Visual Genome: Connecting Language and Vision Using Crowdsourced Dense Image Annotations , 2016, International Journal of Computer Vision.
[3] Xia Hu,et al. Techniques for interpretable machine learning , 2018, Commun. ACM.
[4] Trevor Darrell,et al. Generating Visual Explanations , 2016, ECCV.
[5] Harada Tatsuya,et al. Learning to Explain With Complemental Examples , 2019 .
[6] Trevor Darrell,et al. Textual Explanations for Self-Driving Vehicles , 2018, ECCV.
[7] Trevor Darrell,et al. Generating Counterfactual Explanations with Natural Language , 2018, ICML 2018.
[8] Samy Bengio,et al. Adversarial examples in the physical world , 2016, ICLR.
[9] Andrew Zisserman,et al. Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps , 2013, ICLR.
[10] Bernt Schiele,et al. Evaluation of output embeddings for fine-grained image classification , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[11] Christoph Molnar,et al. Interpretable Machine Learning , 2020 .
[12] Rob Fergus,et al. Visualizing and Understanding Convolutional Networks , 2013, ECCV.
[13] Ananthram Swami,et al. The Limitations of Deep Learning in Adversarial Settings , 2015, 2016 IEEE European Symposium on Security and Privacy (EuroS&P).
[14] Aleksander Madry,et al. Robustness May Be at Odds with Accuracy , 2018, ICLR.
[15] Tatsuya Harada,et al. Learning to Explain With Complemental Examples , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[16] Lei Zhang,et al. Bottom-Up and Top-Down Attention for Image Captioning and Visual Question Answering , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[17] Christoph H. Lampert,et al. Learning to detect unseen object classes by between-class attribute transfer , 2009, 2009 IEEE Conference on Computer Vision and Pattern Recognition.
[18] Pietro Perona,et al. The Caltech-UCSD Birds-200-2011 Dataset , 2011 .
[19] Hang Su,et al. Towards Interpretable Deep Neural Networks by Leveraging Adversarial Examples , 2017, ArXiv.
[20] Ziyan Wu,et al. Counterfactual Visual Explanations , 2019, ICML.
[21] Aleksander Madry,et al. Towards Deep Learning Models Resistant to Adversarial Attacks , 2017, ICLR.
[22] Zhanxing Zhu,et al. Interpreting Adversarially Trained Convolutional Neural Networks , 2019, ICML.
[23] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[24] Shixia Liu,et al. Recent research advances on interactive machine learning , 2018, J. Vis..
[25] Carlos Guestrin,et al. "Why Should I Trust You?": Explaining the Predictions of Any Classifier , 2016, ArXiv.
[26] Trevor Darrell,et al. Multimodal Explanations: Justifying Decisions and Pointing to the Evidence , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[27] Heinz Handels,et al. Interpretable explanations of black box classifiers applied on medical images by meaningful perturbations using variational autoencoders , 2019, Medical Imaging: Image Processing.
[28] Bo Zhang,et al. Improving Interpretability of Deep Neural Networks with Semantic Information , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[29] Ankur Taly,et al. Axiomatic Attribution for Deep Networks , 2017, ICML.
[30] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[31] Alexei A. Efros,et al. Unsupervised Discovery of Mid-Level Discriminative Patches , 2012, ECCV.
[32] Andrea Vedaldi,et al. Interpretable Explanations of Black Boxes by Meaningful Perturbation , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[33] Kouichi Sakurai,et al. One Pixel Attack for Fooling Deep Neural Networks , 2017, IEEE Transactions on Evolutionary Computation.
[34] Bo Zhao,et al. A Large-Scale Attribute Dataset for Zero-Shot Learning , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).
[35] Pradeep Ravikumar,et al. Evaluations and Methods for Explanation through Robustness Analysis , 2019, ICLR.
[36] Trevor Darrell,et al. Grounding Visual Explanations , 2018, ECCV.
[37] Max Welling,et al. Visualizing Deep Neural Network Decisions: Prediction Difference Analysis , 2017, ICLR.
[38] David A. Wagner,et al. Towards Evaluating the Robustness of Neural Networks , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[39] Avanti Shrikumar,et al. Learning Important Features Through Propagating Activation Differences , 2017, ICML.
[40] Seyed-Mohsen Moosavi-Dezfooli,et al. DeepFool: A Simple and Accurate Method to Fool Deep Neural Networks , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).