Embedding Semantic Hierarchy in Discrete Optimal Transport for Risk Minimization

The widely-used cross-entropy (CE) loss-based deep networks achieved significant progress w.r.t. the classification accuracy. However, the CE loss can essentially ignore the risk of misclassification which is usually measured by the distance between the prediction and label in a semantic hierarchical tree. In this paper, we propose to incorporate the risk-aware inter-class correlation in a discrete optimal transport (DOT) training framework by configuring its ground distance matrix. The ground distance matrix can be pre-defined following a priori of hierarchical semantic risk. Specifically, we define the tree induced error (TIE) on a hierarchical semantic tree and extend it to its increasing function from the optimization perspective. The semantic similarity in each level of a tree is integrated with the information gain. We achieve promising results on several large scale image classification tasks with a semantic tree structure in a plug and play manner.

[1]  Marco Cuturi,et al.  Sinkhorn Distances: Lightspeed Computation of Optimal Transport , 2013, NIPS.

[2]  Kibok Lee,et al.  Hierarchical Novelty Detection for Visual Object Recognition , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[3]  Chao Yang,et al.  Dependency-Aware Attention Control for Unconstrained Face Recognition with Image Sets , 2018, ECCV.

[4]  Georgios Paliouras,et al.  Evaluation measures for hierarchical classification: a unified view and novel approaches , 2013, Data Mining and Knowledge Discovery.

[5]  Xiaofeng Liu,et al.  Wasserstein Loss With Alternative Reinforcement Learning for Severity-Aware Semantic Segmentation , 2022, IEEE Transactions on Intelligent Transportation Systems.

[6]  Michelangelo Ceci,et al.  Classifying web documents in a hierarchy of categories: a comprehensive study , 2007, Journal of Intelligent Information Systems.

[7]  Chao Yang,et al.  Ordinal Regression with Neuron Stick-Breaking for Medical Diagnosis , 2018, ECCV Workshops.

[8]  Q. Hu,et al.  Hierarchical Semantic Risk Minimization for Large-Scale Classification , 2021, IEEE Transactions on Cybernetics.

[9]  Xiaofeng Liu,et al.  Severity-Aware Semantic Segmentation With Reinforced Wasserstein Training , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[10]  Alex A. Freitas,et al.  A survey of hierarchical classification across different application domains , 2010, Data Mining and Knowledge Discovery.

[11]  Sean Saito,et al.  HiNet: Hierarchical Classification with Neural Network , 2017, ArXiv.

[12]  G. Griffin,et al.  Caltech-256 Object Category Dataset , 2007 .

[13]  Luc Van Gool,et al.  The Pascal Visual Object Classes (VOC) Challenge , 2010, International Journal of Computer Vision.

[14]  Jianping Fan,et al.  Embedding Visual Hierarchy With Deep Networks for Large-Scale Visual Recognition , 2017, IEEE Transactions on Image Processing.

[15]  Xiaofeng Liu,et al.  Unimodal regularized neuron stick-breaking for ordinal classification , 2020, Neurocomputing.

[16]  Jonathan Krause,et al.  Learning Features and Parts for Fine-Grained Recognition , 2014, 2014 22nd International Conference on Pattern Recognition.

[17]  Hossein Mobahi,et al.  Learning with a Wasserstein Loss , 2015, NIPS.

[18]  Tong Che,et al.  Conservative Wasserstein Training for Pose Estimation , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).

[19]  Qinghua Hu,et al.  Local Bayes Risk Minimization Based Stopping Strategy for Hierarchical Classification , 2017, 2017 IEEE International Conference on Data Mining (ICDM).

[20]  Xiaofeng Liu,et al.  Unimodal-Uniform Constrained Wasserstein Training for Medical Diagnosis , 2019, 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW).

[21]  Ludger Riischendorf The Wasserstein distance and approximation theorems , 1985 .

[22]  Xiaofeng Liu,et al.  Importance-Aware Semantic Segmentation in Self-Driving with Discrete Wasserstein Training , 2020, AAAI.

[23]  Xiaofeng Liu,et al.  Wasserstein Loss based Deep Object Detection , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).

[24]  Mohammed Bennamoun,et al.  Cost-Sensitive Learning of Deep Feature Representations From Imbalanced Data , 2015, IEEE Transactions on Neural Networks and Learning Systems.

[25]  Jonathan Krause,et al.  Hedging your bets: Optimizing accuracy-specificity trade-offs in large scale visual recognition , 2012, 2012 IEEE Conference on Computer Vision and Pattern Recognition.

[26]  Zhengfang Duanmu,et al.  End-to-End Blind Image Quality Assessment Using Deep Neural Networks , 2018, IEEE Transactions on Image Processing.