Towards Visual Explainable Active Learning for Zero-Shot Classification

Zero-shot classification is a promising paradigm to solve an applicable problem when the training classes and test classes are disjoint. Achieving this usually needs experts to externalize their domain knowledge by manually specifying a class-attribute matrix to define which classes have which attributes. Designing a suitable class-attribute matrix is the key to the subsequent procedure, but this design process is tedious and trial-and-error with no guidance. This paper proposes a visual explainable active learning approach with its design and implementation called semantic navigator to solve the above problems. This approach promotes human-AI teaming with four actions (ask, explain, recommend, respond) in each interaction loop. The machine asks contrastive questions to guide humans in the thinking process of attributes. A novel visualization called semantic map explains the current status of the machine. Therefore analysts can better understand why the machine misclassifies objects. Moreover, the machine recommends the labels of classes for each attribute to ease the labeling burden. Finally, humans can steer the model by modifying the labels interactively, and the machine adjusts its recommendations. The visual explainable active learning approach improves humans' efficiency of building zero-shot classification models interactively, compared with the method without guidance. We justify our results with user studies using the standard benchmarks for zero-shot classification.

[1]  Jürgen Bernard,et al.  VIAL: a unified process for visual interactive labeling , 2018, The Visual Computer.

[2]  Zhen Li,et al.  Understanding Hidden Memories of Recurrent Neural Networks , 2017, 2017 IEEE Conference on Visual Analytics Science and Technology (VAST).

[3]  Christopher Andrews,et al.  The human is the loop: new directions for visual analytics , 2014, Journal of Intelligent Information Systems.

[4]  Ning Chen,et al.  Learning Attributes from the Crowdsourced Relative Labels , 2017, AAAI.

[5]  Wei-Lun Chao,et al.  Predicting Visual Exemplars of Unseen Classes for Zero-Shot Learning , 2016, 2017 IEEE International Conference on Computer Vision (ICCV).

[6]  Marco Hutter,et al.  Comparing Visual-Interactive Labeling with Active Learning: An Experimental Study , 2018, IEEE Transactions on Visualization and Computer Graphics.

[7]  Hao Yang,et al.  DQNViz: A Visual Analytics Approach to Understand Deep Q-Networks , 2019, IEEE Transactions on Visualization and Computer Graphics.

[8]  F. Rossi,et al.  The State of the Art in Enhancing Trust in Machine Learning Models with the Use of Visualizations , 2020, Comput. Graph. Forum.

[9]  Christoph H. Lampert,et al.  Zero-Shot Learning—A Comprehensive Evaluation of the Good, the Bad and the Ugly , 2017, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[10]  Pietro Perona,et al.  The Caltech-UCSD Birds-200-2011 Dataset , 2011 .

[11]  Kun Duan,et al.  Discovering localized attributes for fine-grained recognition , 2012, 2012 IEEE Conference on Computer Vision and Pattern Recognition.

[12]  Jacques Bouaud,et al.  Explainable artificial intelligence for breast cancer: A visual case-based reasoning approach , 2019, Artif. Intell. Medicine.

[13]  Shih-Fu Chang,et al.  Designing Category-Level Attributes for Discriminative Visual Recognition , 2013, 2013 IEEE Conference on Computer Vision and Pattern Recognition.

[14]  Gilles Venturini,et al.  VizAssist: an interactive user assistant for visual data mining , 2016, The Visual Computer.

[15]  Christoph H. Lampert,et al.  Learning to detect unseen object classes by between-class attribute transfer , 2009, 2009 IEEE Conference on Computer Vision and Pattern Recognition.

[16]  Shixia Liu,et al.  Recent research advances on interactive machine learning , 2018, J. Vis..

[17]  Chunyan Miao,et al.  A Survey of Zero-Shot Learning , 2019, ACM Trans. Intell. Syst. Technol..

[18]  Jaegul Choo,et al.  Visual Analytics for Explainable Deep Learning , 2018, IEEE Computer Graphics and Applications.

[19]  Kanit Wongsuphasawat,et al.  Voyager: Exploratory Analysis via Faceted Browsing of Visualization Recommendations , 2016, IEEE Transactions on Visualization and Computer Graphics.

[20]  Silvia Miksch,et al.  A Review of Guidance Approaches in Visual Data Analysis: A Multifocal Perspective , 2019, Comput. Graph. Forum.

[21]  Zi Huang,et al.  Leveraging the Invariant Side of Generative Zero-Shot Learning , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[22]  Harmanpreet Kaur,et al.  Building Shared Mental Models between Humans and AI for Effective Collaboration , 2019 .

[23]  P. Cochat,et al.  Et al , 2008, Archives de pediatrie : organe officiel de la Societe francaise de pediatrie.

[24]  Devi Parikh,et al.  Attributes for Classifier Feedback , 2012, ECCV.

[25]  Jiawan Zhang,et al.  Visualizing surrogate decision trees of convolutional neural networks , 2019, J. Vis..

[26]  D. Gentner,et al.  Learning and Transfer: A General Role for Analogical Encoding , 2003 .

[27]  Enrico Bertini,et al.  The Exploratory Labeling Assistant: Mixed-Initiative Label Curation with Large Document Collections , 2018, UIST.

[28]  Alex Endert,et al.  Semantic Interaction for Visual Analytics: Toward Coupling Cognition and Computation , 2014, IEEE Computer Graphics and Applications.

[29]  Burr Settles,et al.  Active Learning Literature Survey , 2009 .

[30]  Ali Farhadi,et al.  Describing objects by their attributes , 2009, 2009 IEEE Conference on Computer Vision and Pattern Recognition.

[31]  Gunther Heidemann,et al.  Inter-active learning of ad-hoc classifiers for video visual analytics , 2012, 2012 IEEE Conference on Visual Analytics Science and Technology (VAST).

[32]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[33]  Martin Wattenberg,et al.  Human-Centered Tools for Coping with Imperfect Algorithms During Medical Decision-Making , 2019, CHI.

[34]  Marco Cavallo,et al.  Clustrophile 2: Guided Visual Clustering Analysis , 2018, IEEE Transactions on Visualization and Computer Graphics.

[35]  Bin Gu,et al.  Scalable Semi-Supervised SVM via Triply Stochastic Gradients , 2019, IJCAI.

[36]  Thomas Ertl,et al.  Visual Classifier Training for Text Document Retrieval , 2012, IEEE Transactions on Visualization and Computer Graphics.

[37]  Rosane Minghim,et al.  An Approach to Supporting Incremental Visual Data Classification , 2015, IEEE Transactions on Visualization and Computer Graphics.

[38]  Yang Chen,et al.  Interactive Correction of Mislabeled Training Data , 2019, 2019 IEEE Conference on Visual Analytics Science and Technology (VAST).

[39]  Laurens van der Maaten,et al.  Learning a Parametric Embedding by Preserving Local Structure , 2009, AISTATS.

[40]  Carla E. Brodley,et al.  Dis-function: Learning distance functions interactively , 2012, 2012 IEEE Conference on Visual Analytics Science and Technology (VAST).

[41]  Daniel,et al.  Default Probability , 2004 .

[42]  Kristen Grauman,et al.  Relative attributes , 2011, 2011 International Conference on Computer Vision.

[43]  Oliver Kramer,et al.  Fast and simple gradient-based optimization for semi-supervised support vector machines , 2014, Neurocomputing.

[44]  Martin Müller,et al.  Towards User‐Centered Active Learning Algorithms , 2018, Comput. Graph. Forum.

[45]  Matthew Berger,et al.  Visually Analyzing and Steering Zero Shot Learning , 2020, 2020 IEEE Visualization Conference (VIS).

[46]  Tao Xiang,et al.  Learning a Deep Embedding Model for Zero-Shot Learning , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[47]  Daniel A. Keim,et al.  SOMFlow: Guided Exploratory Cluster Analysis with Self-Organizing Maps and Analytic Provenance , 2018, IEEE Transactions on Visualization and Computer Graphics.

[48]  David Maxwell Chickering,et al.  Machine Teaching: A New Paradigm for Building Machine Learning Systems , 2017, ArXiv.

[49]  Alexander C. Berg,et al.  Automatic Attribute Discovery and Characterization from Noisy Web Data , 2010, ECCV.

[50]  Minsuk Kahng,et al.  Visual Analytics in Deep Learning: An Interrogative Survey for the Next Frontiers , 2018, IEEE Transactions on Visualization and Computer Graphics.

[51]  Jing Wu,et al.  Visual Diagnosis of Tree Boosting Methods , 2018, IEEE Transactions on Visualization and Computer Graphics.

[52]  Alexander M. Rush,et al.  LSTMVis: A Tool for Visual Analysis of Hidden State Dynamics in Recurrent Neural Networks , 2016, IEEE Transactions on Visualization and Computer Graphics.

[53]  Kristen Grauman,et al.  Interactively building a discriminative vocabulary of nameable attributes , 2011, CVPR 2011.

[54]  Chris North,et al.  SIRIUS: Dual, Symmetric, Interactive Dimension Reductions , 2019, IEEE Transactions on Visualization and Computer Graphics.

[55]  Chris North,et al.  Semantic Interaction for Sensemaking: Inferring Analytical Reasoning for Model Steering , 2012, IEEE Transactions on Visualization and Computer Graphics.

[56]  Aditya G. Parameswaran,et al.  SeeDB: Efficient Data-Driven Visualization Recommendations to Support Visual Analytics , 2015, Proc. VLDB Endow..

[57]  Zhen Li,et al.  Towards Better Analysis of Deep Convolutional Neural Networks , 2016, IEEE Transactions on Visualization and Computer Graphics.

[58]  Sheng Tang,et al.  Zero-Shot Learning With Attribute Selection , 2018, AAAI.

[59]  Huamin Qu,et al.  RuleMatrix: Visualizing and Understanding Classifiers with Rules , 2018, IEEE Transactions on Visualization and Computer Graphics.

[60]  Jerry Alan Fails,et al.  Interactive machine learning , 2003, IUI '03.

[61]  Tim Miller,et al.  Explanation in Artificial Intelligence: Insights from the Social Sciences , 2017, Artif. Intell..

[62]  Jun Zhu,et al.  Analyzing the Training Processes of Deep Generative Models , 2018, IEEE Transactions on Visualization and Computer Graphics.

[63]  Edward Lank,et al.  Label-and-Learn: Visualizing the Likelihood of Machine Learning Classifier's Success During Data Labeling , 2017, IUI.

[64]  Huamin Qu,et al.  ProtoSteer: Steering Deep Sequence Model with Prototypes , 2020, IEEE Transactions on Visualization and Computer Graphics.

[65]  Michael Granitzer,et al.  User-Based Active Learning , 2010, 2010 IEEE International Conference on Data Mining Workshops.

[66]  Aaron Hertzmann,et al.  DesignScape: Design with Interactive Layout Suggestions , 2015, CHI.

[67]  Claudia Müller-Birn,et al.  Coordinating Agents: Promoting Shared Situational Awareness in Collaborative Sensemaking , 2018, CSCW Companion.

[68]  Changjian Chen,et al.  An Interactive Method to Improve Crowdsourced Annotations , 2019, IEEE Transactions on Visualization and Computer Graphics.

[69]  Adriana Kovashka,et al.  Relative Attributes for Enhanced Human-Machine Communication , 2012, AAAI.

[70]  Ayhan Demiriz,et al.  Semi-Supervised Support Vector Machines , 1998, NIPS.

[71]  Trevor Darrell,et al.  Generalized Zero- and Few-Shot Learning via Aligned Variational Autoencoders , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[72]  Geoffrey E. Hinton,et al.  Visualizing Data using t-SNE , 2008 .

[73]  James Zou,et al.  Towards Automatic Concept-based Explanations , 2019, NeurIPS.

[74]  Philip H. S. Torr,et al.  An embarrassingly simple approach to zero-shot learning , 2015, ICML.