OCTET: Object-aware Counterfactual Explanations

Nowadays, deep vision models are being widely deployed in safety-critical applications, e.g., autonomous driving, and explainability of such models is becoming a pressing concern. Among explanation methods, counterfactual explanations aim to find minimal and interpretable changes to the input image that would also change the output of the model to be explained. Such explanations point end-users at the main factors that impact the decision of the model. However, previous methods struggle to explain decision models trained on images with many objects, e.g., urban scenes, which are more difficult to work with but also arguably more critical to explain. In this work, we propose to tackle this issue with an object-centric framework for counterfactual explanation generation. Our method, inspired by recent generative modeling works, encodes the query image into a latent space that is structured in a way to ease object-level manipulations. Doing so, it provides the end-user with control over which search directions (e.g., spatial displacement of objects, style modification, etc.) are to be explored during the counterfactual generation. We conduct a set of experiments on counterfactual explanation benchmarks for driving scenes, and we show that our method can be adapted beyond classification, e.g., to explain semantic segmentation models. To complete our analysis, we design and run a user study that measures the usefulness of counterfactual explanations in understanding a decision model. Code is available at https://github.com/valeoai/OCTET.

[1]  Matthias Hein,et al.  Diffusion Visual Counterfactual Explanations , 2022, NeurIPS.

[2]  Alexei A. Efros,et al.  BlobGAN: Spatially Disentangled Scene Representations , 2022, ECCV.

[3]  F. Jurie,et al.  Diffusion Models for Counterfactual Explanations , 2022, ACCV.

[4]  Li Fuxin,et al.  Cycle-Consistent Counterfactuals by Latent Transformations , 2022, 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[5]  D. Mahajan,et al.  Making Heads or Tails: Towards Semantically Consistent Visual Counterfactuals , 2022, ECCV.

[6]  Thomas Serre,et al.  What I Cannot Predict, I Do Not Understand: A Human-Centered Evaluation Framework for Explainability Methods , 2021, NeurIPS.

[7]  M. Cord,et al.  STEEX: Steering Counterfactual Explanations with Semantics , 2021, ECCV.

[8]  P. Pérez,et al.  Explainability of Deep Vision-Based Autonomous Driving Systems: Review and Challenges , 2021, International Journal of Computer Vision.

[9]  David Martens,et al.  Explainable Image Classification with Evidence Counterfactual , 2020, ArXiv.

[10]  Anh M Nguyen,et al.  The effectiveness of feature attribution methods and its correlation with automatic evaluation scores , 2021, NeurIPS.

[11]  Prafulla Dhariwal,et al.  Diffusion Models Beat GANs on Image Synthesis , 2021, NeurIPS.

[12]  Chenliang Xu,et al.  Discover the Unknown Biased Attribute of an Image Classifier , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).

[13]  Avinatan Hassidim,et al.  Explaining in Style: Training a GAN to explain a classifier in StyleSpace , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).

[14]  Alexandre Lacoste,et al.  Beyond Trivial Counterfactual Explanations with Diverse Valuable Explanations , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).

[15]  Daniel Cohen-Or,et al.  Designing an encoder for StyleGAN image manipulation , 2021, ACM Trans. Graph..

[16]  Andreas Geiger,et al.  GIRAFFE: Representing Scenes as Compositional Generative Neural Feature Fields , 2020, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[17]  Daniel Cohen-Or,et al.  Encoding in Style: a StyleGAN Encoder for Image-to-Image Translation , 2020, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[18]  Cuntai Guan,et al.  A Survey on Explainable Artificial Intelligence (XAI): Toward Medical XAI , 2019, IEEE Transactions on Neural Networks and Learning Systems.

[19]  Chirag Agarwal,et al.  On the Connections between Counterfactual Explanations and Adversarial Examples , 2021, ArXiv.

[20]  Ben Swift,et al.  Semantics and explanation: why counterfactual explanations produce adversarial examples in deep neural networks , 2020, ArXiv.

[21]  John P. Dickerson,et al.  Counterfactual Explanations for Machine Learning: A Review , 2020, ArXiv.

[22]  Timo Freiesleben,et al.  Counterfactual Explanations & Adversarial Examples - Common Grounds, Essential Differences, and Potential Transfers , 2020, ArXiv.

[23]  Katherine Driggs Campbell,et al.  To Explain or Not to Explain: A Study on the Necessity of Explanations for Autonomous Vehicles , 2020, ArXiv.

[24]  Mohit Bansal,et al.  Evaluating Explainable AI: Which Algorithmic Explanations Help Users Predict Model Behavior? , 2020, ACL.

[25]  Nuno Vasconcelos,et al.  SCOUT: Self-Aware Discriminant Counterfactual Explanations , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[26]  Andrea Vedaldi,et al.  There and Back Again: Revisiting Backpropagation Saliency Methods , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[27]  Bolei Zhou,et al.  In-Domain GAN Inversion for Real Image Editing , 2020, ECCV.

[28]  Nuno Vasconcelos,et al.  Explainable Object-Induced Action Decision for Autonomous Vehicles , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[29]  X. Jessie Yang,et al.  Expectations and Trust in Automated Vehicles , 2020, CHI Extended Abstracts.

[30]  Peter Wonka,et al.  SEAN: Image Synthesis With Semantic Region-Adaptive Normalization , 2019, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[31]  Peter Wonka,et al.  Image2StyleGAN++: How to Edit the Embedded Images? , 2019, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[32]  K. Batmanghelich,et al.  Explanation by Progressive Exaggeration , 2019, ICLR.

[33]  Trevor Darrell,et al.  BDD100K: A Diverse Driving Dataset for Heterogeneous Multitask Learning , 2018, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[34]  Anna Khoreva,et al.  Grid Saliency for Context Explanations of Semantic Segmentation , 2019, NeurIPS.

[35]  Sven Behnke,et al.  Interpretable and Fine-Grained Visual Explanations for Convolutional Neural Networks , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[36]  Ziyan Wu,et al.  Counterfactual Visual Explanations , 2019, ICML.

[37]  Peter Wonka,et al.  Image2StyleGAN: How to Embed Images Into the StyleGAN Latent Space? , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).

[38]  David Duvenaud,et al.  Explaining Image Classifiers by Counterfactual Generation , 2018, ICLR.

[39]  Cynthia Rudin,et al.  This Looks Like That: Deep Learning for Interpretable Image Recognition , 2018 .

[40]  Been Kim,et al.  Sanity Checks for Saliency Maps , 2018, NeurIPS.

[41]  Mariusz Bojarski,et al.  VisualBackProp: Efficient Visualization of CNNs for Autonomous Driving , 2018, 2018 IEEE International Conference on Robotics and Automation (ICRA).

[42]  Trevor Darrell,et al.  Grounding Visual Explanations , 2018, ECCV.

[43]  Andreas Geiger,et al.  Conditional Affordance Learning for Driving in Urban Environments , 2018, CoRL.

[44]  Roger B. Grosse,et al.  Isolating Sources of Disentanglement in Variational Autoencoders , 2018, NeurIPS.

[45]  Brian E. Ruttenberg,et al.  Causal Learning and Explanation of Deep Neural Networks via Autoencoded Activations , 2018, ArXiv.

[46]  Alexei A. Efros,et al.  The Unreasonable Effectiveness of Deep Features as a Perceptual Metric , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[47]  Quanshi Zhang,et al.  Interpretable Convolutional Neural Networks , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[48]  Geoffrey E. Hinton,et al.  Distilling a Neural Network Into a Soft Decision Tree , 2017, CEx@AI*IA.

[49]  Chris Russell,et al.  Counterfactual Explanations Without Opening the Black Box: Automated Decisions and the GDPR , 2017, ArXiv.

[50]  Sepp Hochreiter,et al.  GANs Trained by a Two Time-Scale Update Rule Converge to a Local Nash Equilibrium , 2017, NIPS.

[51]  George Papandreou,et al.  Rethinking Atrous Convolution for Semantic Image Segmentation , 2017, ArXiv.

[52]  Andrea Vedaldi,et al.  Interpretable Explanations of Black Boxes by Meaningful Perturbation , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[53]  John F. Canny,et al.  Interpretable Learning for Self-Driving Cars by Visualizing Causal Attention , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[54]  Ankur Taly,et al.  Axiomatic Attribution for Deep Networks , 2017, ICML.

[55]  Been Kim,et al.  Towards A Rigorous Science of Interpretable Machine Learning , 2017, 1702.08608.

[56]  Abhishek Das,et al.  Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization , 2016, 2017 IEEE International Conference on Computer Vision (ICCV).

[57]  Kilian Q. Weinberger,et al.  Densely Connected Convolutional Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[58]  Carlos Guestrin,et al.  "Why Should I Trust You?": Explaining the Predictions of Any Classifier , 2016, ArXiv.

[59]  Seyed-Mohsen Moosavi-Dezfooli,et al.  DeepFool: A Simple and Accurate Method to Fool Deep Neural Networks , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[60]  Oluwasanmi Koyejo,et al.  Examples are not enough, learn to criticize! Criticism for Interpretability , 2016, NIPS.

[61]  Yinda Zhang,et al.  LSUN: Construction of a Large-scale Image Dataset using Deep Learning with Humans in the Loop , 2015, ArXiv.

[62]  Bolei Zhou,et al.  Object Detectors Emerge in Deep Scene CNNs , 2014, ICLR.

[63]  Jonathon Shlens,et al.  Explaining and Harnessing Adversarial Examples , 2014, ICLR.

[64]  Xiaogang Wang,et al.  Deep Learning Face Attributes in the Wild , 2014, 2015 IEEE International Conference on Computer Vision (ICCV).

[65]  Joan Bruna,et al.  Intriguing properties of neural networks , 2013, ICLR.

[66]  Rob Fergus,et al.  Visualizing and Understanding Convolutional Networks , 2013, ECCV.

[67]  Pietro Perona,et al.  The Caltech-UCSD Birds-200-2011 Dataset , 2011 .

[68]  Fei-Fei Li,et al.  ImageNet: A large-scale hierarchical image database , 2009, 2009 IEEE Conference on Computer Vision and Pattern Recognition.

[69]  Leo P. Kadanoff,et al.  The Unreasonable Effectiveness of , 2000 .