OCTET: Object-aware Counterfactual Explanations
暂无分享,去创建一个
[1] Matthias Hein,et al. Diffusion Visual Counterfactual Explanations , 2022, NeurIPS.
[2] Alexei A. Efros,et al. BlobGAN: Spatially Disentangled Scene Representations , 2022, ECCV.
[3] F. Jurie,et al. Diffusion Models for Counterfactual Explanations , 2022, ACCV.
[4] Li Fuxin,et al. Cycle-Consistent Counterfactuals by Latent Transformations , 2022, 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[5] D. Mahajan,et al. Making Heads or Tails: Towards Semantically Consistent Visual Counterfactuals , 2022, ECCV.
[6] Thomas Serre,et al. What I Cannot Predict, I Do Not Understand: A Human-Centered Evaluation Framework for Explainability Methods , 2021, NeurIPS.
[7] M. Cord,et al. STEEX: Steering Counterfactual Explanations with Semantics , 2021, ECCV.
[8] P. Pérez,et al. Explainability of Deep Vision-Based Autonomous Driving Systems: Review and Challenges , 2021, International Journal of Computer Vision.
[9] David Martens,et al. Explainable Image Classification with Evidence Counterfactual , 2020, ArXiv.
[10] Anh M Nguyen,et al. The effectiveness of feature attribution methods and its correlation with automatic evaluation scores , 2021, NeurIPS.
[11] Prafulla Dhariwal,et al. Diffusion Models Beat GANs on Image Synthesis , 2021, NeurIPS.
[12] Chenliang Xu,et al. Discover the Unknown Biased Attribute of an Image Classifier , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).
[13] Avinatan Hassidim,et al. Explaining in Style: Training a GAN to explain a classifier in StyleSpace , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).
[14] Alexandre Lacoste,et al. Beyond Trivial Counterfactual Explanations with Diverse Valuable Explanations , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).
[15] Daniel Cohen-Or,et al. Designing an encoder for StyleGAN image manipulation , 2021, ACM Trans. Graph..
[16] Andreas Geiger,et al. GIRAFFE: Representing Scenes as Compositional Generative Neural Feature Fields , 2020, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[17] Daniel Cohen-Or,et al. Encoding in Style: a StyleGAN Encoder for Image-to-Image Translation , 2020, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[18] Cuntai Guan,et al. A Survey on Explainable Artificial Intelligence (XAI): Toward Medical XAI , 2019, IEEE Transactions on Neural Networks and Learning Systems.
[19] Chirag Agarwal,et al. On the Connections between Counterfactual Explanations and Adversarial Examples , 2021, ArXiv.
[20] Ben Swift,et al. Semantics and explanation: why counterfactual explanations produce adversarial examples in deep neural networks , 2020, ArXiv.
[21] John P. Dickerson,et al. Counterfactual Explanations for Machine Learning: A Review , 2020, ArXiv.
[22] Timo Freiesleben,et al. Counterfactual Explanations & Adversarial Examples - Common Grounds, Essential Differences, and Potential Transfers , 2020, ArXiv.
[23] Katherine Driggs Campbell,et al. To Explain or Not to Explain: A Study on the Necessity of Explanations for Autonomous Vehicles , 2020, ArXiv.
[24] Mohit Bansal,et al. Evaluating Explainable AI: Which Algorithmic Explanations Help Users Predict Model Behavior? , 2020, ACL.
[25] Nuno Vasconcelos,et al. SCOUT: Self-Aware Discriminant Counterfactual Explanations , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[26] Andrea Vedaldi,et al. There and Back Again: Revisiting Backpropagation Saliency Methods , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[27] Bolei Zhou,et al. In-Domain GAN Inversion for Real Image Editing , 2020, ECCV.
[28] Nuno Vasconcelos,et al. Explainable Object-Induced Action Decision for Autonomous Vehicles , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[29] X. Jessie Yang,et al. Expectations and Trust in Automated Vehicles , 2020, CHI Extended Abstracts.
[30] Peter Wonka,et al. SEAN: Image Synthesis With Semantic Region-Adaptive Normalization , 2019, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[31] Peter Wonka,et al. Image2StyleGAN++: How to Edit the Embedded Images? , 2019, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[32] K. Batmanghelich,et al. Explanation by Progressive Exaggeration , 2019, ICLR.
[33] Trevor Darrell,et al. BDD100K: A Diverse Driving Dataset for Heterogeneous Multitask Learning , 2018, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[34] Anna Khoreva,et al. Grid Saliency for Context Explanations of Semantic Segmentation , 2019, NeurIPS.
[35] Sven Behnke,et al. Interpretable and Fine-Grained Visual Explanations for Convolutional Neural Networks , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[36] Ziyan Wu,et al. Counterfactual Visual Explanations , 2019, ICML.
[37] Peter Wonka,et al. Image2StyleGAN: How to Embed Images Into the StyleGAN Latent Space? , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[38] David Duvenaud,et al. Explaining Image Classifiers by Counterfactual Generation , 2018, ICLR.
[39] Cynthia Rudin,et al. This Looks Like That: Deep Learning for Interpretable Image Recognition , 2018 .
[40] Been Kim,et al. Sanity Checks for Saliency Maps , 2018, NeurIPS.
[41] Mariusz Bojarski,et al. VisualBackProp: Efficient Visualization of CNNs for Autonomous Driving , 2018, 2018 IEEE International Conference on Robotics and Automation (ICRA).
[42] Trevor Darrell,et al. Grounding Visual Explanations , 2018, ECCV.
[43] Andreas Geiger,et al. Conditional Affordance Learning for Driving in Urban Environments , 2018, CoRL.
[44] Roger B. Grosse,et al. Isolating Sources of Disentanglement in Variational Autoencoders , 2018, NeurIPS.
[45] Brian E. Ruttenberg,et al. Causal Learning and Explanation of Deep Neural Networks via Autoencoded Activations , 2018, ArXiv.
[46] Alexei A. Efros,et al. The Unreasonable Effectiveness of Deep Features as a Perceptual Metric , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[47] Quanshi Zhang,et al. Interpretable Convolutional Neural Networks , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[48] Geoffrey E. Hinton,et al. Distilling a Neural Network Into a Soft Decision Tree , 2017, CEx@AI*IA.
[49] Chris Russell,et al. Counterfactual Explanations Without Opening the Black Box: Automated Decisions and the GDPR , 2017, ArXiv.
[50] Sepp Hochreiter,et al. GANs Trained by a Two Time-Scale Update Rule Converge to a Local Nash Equilibrium , 2017, NIPS.
[51] George Papandreou,et al. Rethinking Atrous Convolution for Semantic Image Segmentation , 2017, ArXiv.
[52] Andrea Vedaldi,et al. Interpretable Explanations of Black Boxes by Meaningful Perturbation , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[53] John F. Canny,et al. Interpretable Learning for Self-Driving Cars by Visualizing Causal Attention , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[54] Ankur Taly,et al. Axiomatic Attribution for Deep Networks , 2017, ICML.
[55] Been Kim,et al. Towards A Rigorous Science of Interpretable Machine Learning , 2017, 1702.08608.
[56] Abhishek Das,et al. Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization , 2016, 2017 IEEE International Conference on Computer Vision (ICCV).
[57] Kilian Q. Weinberger,et al. Densely Connected Convolutional Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[58] Carlos Guestrin,et al. "Why Should I Trust You?": Explaining the Predictions of Any Classifier , 2016, ArXiv.
[59] Seyed-Mohsen Moosavi-Dezfooli,et al. DeepFool: A Simple and Accurate Method to Fool Deep Neural Networks , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[60] Oluwasanmi Koyejo,et al. Examples are not enough, learn to criticize! Criticism for Interpretability , 2016, NIPS.
[61] Yinda Zhang,et al. LSUN: Construction of a Large-scale Image Dataset using Deep Learning with Humans in the Loop , 2015, ArXiv.
[62] Bolei Zhou,et al. Object Detectors Emerge in Deep Scene CNNs , 2014, ICLR.
[63] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[64] Xiaogang Wang,et al. Deep Learning Face Attributes in the Wild , 2014, 2015 IEEE International Conference on Computer Vision (ICCV).
[65] Joan Bruna,et al. Intriguing properties of neural networks , 2013, ICLR.
[66] Rob Fergus,et al. Visualizing and Understanding Convolutional Networks , 2013, ECCV.
[67] Pietro Perona,et al. The Caltech-UCSD Birds-200-2011 Dataset , 2011 .
[68] Fei-Fei Li,et al. ImageNet: A large-scale hierarchical image database , 2009, 2009 IEEE Conference on Computer Vision and Pattern Recognition.
[69] Leo P. Kadanoff,et al. The Unreasonable Effectiveness of , 2000 .