Feature and Label Association Based on Granulation Entropy for Deep Neural Networks

Pooling layers help reduce redundancy and the number of parameters before building a multilayered neural network that performs the remaining processing operations. Usually, pooling operators in deep learning models use an explicit topological organization, which is not always possible to obtain on multi-label data. In a previous paper, we proposed a pooling architecture based on association to deal with this issue. The association was defined by means of Pearson’s correlation. However, features must exhibit a certain degree of correlation with each other, which might not hold in all situations. In this paper, we propose a new method that replaces the correlation measure with another one that computes the entropy in the information granules that are generated from two features or labels. Numerical simulations have shown that our proposal is superior in those datasets with low correlation. This means that it induces a significant reduction in the number of parameters of neural networks, without affecting their accuracy.

[1]  Shengcai Liao,et al.  Multi-label convolutional neural network based pedestrian attribute classification , 2017, Image Vis. Comput..

[2]  Yiyu Yao,et al.  Probabilistic approaches to rough sets , 2003, Expert Syst. J. Knowl. Eng..

[3]  Francisco Charte,et al.  Multilabel Classification , 2016, Springer International Publishing.

[4]  Zhuowen Tu,et al.  Generalizing Pooling Functions in Convolutional Neural Networks: Mixed, Gated, and Tree , 2015, AISTATS.

[5]  Min-Ling Zhang,et al.  A Review on Multi-Label Learning Algorithms , 2014, IEEE Transactions on Knowledge and Data Engineering.

[6]  Bingbing Ni,et al.  HCP: A Flexible CNN Framework for Multi-Label Image Classification , 2014, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[7]  Geoffrey E. Hinton,et al.  Reducing the Dimensionality of Data with Neural Networks , 2006, Science.

[8]  Witold Pedrycz,et al.  Granular Computing: At the Junction of Rough Sets and Fuzzy Sets , 2008 .

[9]  Theresa Beaubouef,et al.  Rough Sets , 2019, Lecture Notes in Computer Science.

[10]  Ramakanth Kavuluru,et al.  Convolutional neural networks for biomedical text classification: application in indexing biomedical articles , 2015, BCB.

[11]  Qiang Chen,et al.  Network In Network , 2013, ICLR.

[12]  Mark B. Sandler,et al.  Automatic Tagging Using Deep Convolutional Neural Networks , 2016, ISMIR.

[13]  Urszula Wybraniec-Skardowska,et al.  Extensions and Intentions in the Ruogh Set Theory , 1998, Inf. Sci..

[14]  Vladik Kreinovich,et al.  Handbook of Granular Computing , 2008 .

[15]  Daniel Vanderpooten,et al.  A Generalized Definition of Rough Approximations Based on Similarity , 2000, IEEE Trans. Knowl. Data Eng..

[16]  Francisco Charte,et al.  Multilabel Classification: Problem Analysis, Metrics and Techniques , 2016 .

[17]  Guigang Zhang,et al.  Deep Learning , 2016, Int. J. Semantic Comput..

[18]  Francisco Charte,et al.  R Ultimate Multilabel Dataset Repository , 2016, HAIS.

[19]  Yiyu Yao,et al.  Induction of Classification Rules by Granular Computing , 2002, Rough Sets and Current Trends in Computing.

[20]  Jingtao Yao,et al.  A granular computing framework for self-organizing maps , 2009, Neurocomputing.

[21]  Koen Vanhoof,et al.  Deep neural network to extract high-level features and labels in multi-label classification problems , 2020, Neurocomputing.