Multi-Attribute Learning With Highly Imbalanced Data

Data is one of the most important keys for success when studying a simple or a complex phenomenon. With the use of deep learning exploding and its democratization, non-computer science experts may struggle to use highly complex deep learning architectures, even when straightforward models offer them suitable performances. In this article, we study the specific and common problem of data imbalance in real databases as most of the bad performance problems are due to the data itself. We cover two keys aspects. First, we propose ways to deal with the situation when a data set contains different levels of imbalance. Classical imbalanced learning strategies cannot be directly applied when using multi-attribute deep learning models, i.e., multi-task or multi-label architectures. Therefore, one of our contributions is a proposed adaptation to face each one of the problems derived from imbalance. Second, we demonstrate that with little to no imbalance, straightforward deep learning models work well. However, for non-experts, these models can be seen as black boxes, where all efforts are invested in preprocessing the data. To simplify the problem, we perform the classification task without features that are costly to extract, such as part localization which is widely used in the state of the art of attribute classification. We make use of three widely known attribute databases, CUB-200-2011 - CUB as our main use case due to its deeply imbalanced nature, along with two balanced databases: celebA and AwA2. All of them contain multi-attribute annotations. The results of very fine-grained attribute learning demonstrate that in the presence of imbalance, our proposed strategies make it possible to have competitive results against the state of the art while taking advantage of multi-attribute deep learning models. We also noticed an increase in performance while using a specialized loss function (Focal Loss). For CUB, we have competitive results, and for CelebA and AwA2 our strategies over-perform the state of the art.

[1]  Shiguang Shan,et al.  Multiset Feature Learning for Highly Imbalanced Data Classification , 2017, AAAI.

[2]  Yuhong Guo,et al.  Zero-Shot Classification with Discriminative Semantic Representation Learning , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[3]  Feiyue Huang,et al.  Attention Convolutional Binary Neural Tree for Fine-Grained Visual Categorization , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[4]  Yang Song,et al.  Class-Balanced Loss Based on Effective Number of Samples , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[5]  Leo Breiman,et al.  Random Forests , 2001, Machine Learning.

[6]  Ying Wu,et al.  Recognizing Part Attributes With Insufficient Data , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).

[7]  Chen Sun,et al.  Revisiting Unreasonable Effectiveness of Data in Deep Learning Era , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[8]  Wei Wu,et al.  Dynamic Curriculum Learning for Imbalanced Data Classification , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).

[9]  Pietro Perona,et al.  The Caltech-UCSD Birds-200-2011 Dataset , 2011 .

[10]  Bharath Hariharan,et al.  Revisiting Pose-Normalization for Fine-Grained Few-Shot Recognition , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[11]  G. Zhai,et al.  Three-branch and Mutil-scale learning for Fine-grained Image Recognition (TBMSL-Net) , 2020, ArXiv.

[12]  Zhi-Hua Zhou,et al.  ML-KNN: A lazy learning approach to multi-label learning , 2007, Pattern Recognit..

[13]  Haojie Li,et al.  Progressive learning for weakly supervised fine-grained classification , 2020, Signal Process..

[14]  Wei Wu,et al.  Hierarchical Feature Embedding for Attribute Recognition , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[15]  Xiaogang Wang,et al.  Deep Learning Face Attributes in the Wild , 2014, 2015 IEEE International Conference on Computer Vision (ICCV).

[16]  Hemant Ishwaran,et al.  A random forests quantile classifier for class imbalanced data , 2019, Pattern Recognit..

[17]  Pedro Antonio Gutiérrez,et al.  MELM-GRBF: A modified version of the extreme learning machine for generalized radial basis function neural networks , 2011, Neurocomputing.

[18]  Ross B. Girshick,et al.  Focal Loss for Dense Object Detection , 2017, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[19]  Yijing Li,et al.  Learning from class-imbalanced data: Review of methods and applications , 2017, Expert Syst. Appl..

[20]  G. Golub,et al.  Updating formulae and a pairwise algorithm for computing sample variances , 1979 .

[21]  Raphaël Marée,et al.  Fast Multi-class Image Annotation with Random Subwindows and Multiple Output Randomized Trees , 2009, VISAPP.

[22]  Dat T. Huynh,et al.  Fine-Grained Generalized Zero-Shot Learning via Dense Attribute-Based Attention , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[23]  Xiao Liu,et al.  Localizing by Describing: Attribute-Guided Attention Localization for Fine-Grained Recognition , 2016, AAAI.

[24]  Bernt Schiele,et al.  Feature Generating Networks for Zero-Shot Learning , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[25]  Nikolaos Doulamis,et al.  Deep Learning for Computer Vision: A Brief Review , 2018, Comput. Intell. Neurosci..

[26]  Zhongfei Zhang,et al.  A Semantics-Guided Class Imbalance Learning Model for Zero-Shot Classification , 2019, ArXiv.

[27]  Cordelia Schmid,et al.  Label-Embedding for Attribute-Based Classification , 2013, 2013 IEEE Conference on Computer Vision and Pattern Recognition.

[28]  Longbing Cao,et al.  Training deep neural networks on imbalanced data sets , 2016, 2016 International Joint Conference on Neural Networks (IJCNN).

[29]  Fatemeh Afsari,et al.  Hesitant fuzzy decision tree approach for highly imbalanced data classification , 2017, Appl. Soft Comput..

[30]  Chen Huang,et al.  Learning Deep Representation for Imbalanced Classification , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[31]  Ioannis A. Kakadiaris,et al.  Deep Imbalanced Attribute Classification using Visual Attention Aggregation , 2018, ECCV.

[32]  Trevor Darrell,et al.  PANDA: Pose Aligned Networks for Deep Attribute Modeling , 2013, 2014 IEEE Conference on Computer Vision and Pattern Recognition.

[33]  Shaogang Gong,et al.  Class Rectification Hard Mining for Imbalanced Deep Learning , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[34]  Chen Huang,et al.  Deep Imbalanced Learning for Face Recognition and Attribute Prediction , 2018, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[35]  Nazli Ikizler-Cinbis,et al.  Attributes2Classname: A Discriminative Model for Attribute-Based Unsupervised Zero-Shot Learning , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[36]  Gal Chechik,et al.  Long-tail learning with attributes , 2020, ArXiv.

[37]  Andrew Zisserman,et al.  Learning Visual Attributes , 2007, NIPS.

[38]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).