ALDL: a novel method for label distribution learning

AbstractData complexity has increased manifold in the age of data-driven societies. The data has become huge and inherently complex. The single-label classification algorithms that were discrete in their operation are losing prominence since the nature of data is not monolithic anymore. There are now cases in machine learning where data may belong to more than one class or multiple classes. This nature of data has created the need for new algorithms or methods that are multi-label in nature. Label distribution learning (LDL) is a new way to view multi-labelled algorithms. It tries to quantify the degree to which a label defines an instance. Therefore, for every instance there is a label distribution. In this paper, we introduce a new learning method, namely, angular label distribution learning (ALDL). It is based on the angular distribution function, which is derived from the computation of the length of the arc connecting two points in a circle. Comparative performance evaluation in terms of mean-square error (MSE) of the proposed ALDL has been made with algorithm adaptation of k-NN (AA-kNN), multilayer perceptron, Levenberg–Marquardt neural network and layer-recurrent neural network LDL datasets. MSE is observed to decrease for the proposed ALDL. ALDL is also highly statistically significant for the real world datasets when compared with the standard algorithms for LDL.

[1]  Bidyut Baran Chaudhuri,et al.  Efficient training and improved performance of multilayer perceptron in pattern classification , 2000, Neurocomputing.

[2]  D. Raab,et al.  A cosine approximation to the normal distribution , 1961 .

[3]  Martin T. Hagan,et al.  Neural network design , 1995 .

[4]  Zhi-Hua Zhou,et al.  Multi-instance multi-label learning , 2008, Artif. Intell..

[5]  Grigorios Tsoumakas,et al.  Random K-labelsets for Multilabel Classification , 2022 .

[6]  E. Parzen On Estimation of a Probability Density Function and Mode , 1962 .

[7]  Kate Smith-Miles,et al.  Facial age estimation by multilinear subspace analysis , 2009, 2009 IEEE International Conference on Acoustics, Speech and Signal Processing.

[8]  Grigorios Tsoumakas,et al.  A Triple-Random Ensemble Classification Method for Mining Multi-label Data , 2010, 2010 IEEE International Conference on Data Mining Workshops.

[9]  G. A. Marcoulides,et al.  Discovering Knowledge in Data: an Introduction to Data Mining , 2005 .

[10]  D. Botstein,et al.  Cluster analysis and display of genome-wide expression patterns. , 1998, Proceedings of the National Academy of Sciences of the United States of America.

[11]  Evangelos E. Milios,et al.  Narrative text classification for automatic key phrase extraction in web document corpora , 2005, WIDM '05.

[12]  Terence D. Sanger,et al.  Optimal unsupervised learning in a single-layer linear feedforward neural network , 1989, Neural Networks.

[13]  Daniel T. Larose,et al.  Discovering Knowledge in Data: An Introduction to Data Mining , 2005 .

[14]  T. Cacoullos Estimation of a multivariate density , 1966 .

[15]  Michael J. Lyons,et al.  Coding facial expressions with Gabor wavelets , 1998, Proceedings Third IEEE International Conference on Automatic Face and Gesture Recognition.

[16]  Xin Geng,et al.  Label Distribution Learning , 2013, 2013 IEEE 13th International Conference on Data Mining Workshops.

[17]  Vladimir Vapnik,et al.  Statistical learning theory , 1998 .

[18]  Jiebo Luo,et al.  Learning multi-label scene classification , 2004, Pattern Recognit..

[19]  Zhi-Hua Zhou,et al.  ML-KNN: A lazy learning approach to multi-label learning , 2007, Pattern Recognit..

[20]  Geoff Holmes,et al.  Multi-label Classification Using Ensembles of Pruned Sets , 2008, 2008 Eighth IEEE International Conference on Data Mining.

[21]  Matti Pietikäinen,et al.  Face Description with Local Binary Patterns: Application to Face Recognition , 2006, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[22]  M. R. Leadbetter,et al.  On the Estimation of the Probability Density, I , 1963 .

[23]  Jiawei Han,et al.  Data Mining: Concepts and Techniques , 2000 .

[24]  Yihong Gong,et al.  Multi-labelled classification using maximum entropy method , 2005, SIGIR '05.

[25]  Qingshan Liu,et al.  A One-Layer Recurrent Neural Network With a Discontinuous Hard-Limiting Activation Function for Quadratic Programming , 2008, IEEE Transactions on Neural Networks.

[26]  Xin Geng,et al.  Multilabel Ranking with Inconsistent Rankers , 2014, 2014 IEEE Conference on Computer Vision and Pattern Recognition.

[27]  M. Pravin Kumar,et al.  BACKPROPAGATION LEARNING ALGORITHM BASED ON LEVENBERG MARQUARDT ALGORITHM , 2012 .

[28]  Grigorios Tsoumakas,et al.  Multi-Label Classification: An Overview , 2007, Int. J. Data Warehous. Min..

[29]  Zhi-Hua Zhou,et al.  A k-nearest neighbor based algorithm for multi-label classification , 2005, 2005 IEEE International Conference on Granular Computing.

[30]  Yassine Ben Ayed,et al.  Hybrid SVM/HMM model for the arab phonemes recognition , 2016, Int. Arab J. Inf. Technol..

[31]  Zhi-Hua Zhou,et al.  Facial Age Estimation by Learning from Label Distributions , 2010, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[32]  Tin Kam Ho,et al.  The Random Subspace Method for Constructing Decision Forests , 1998, IEEE Trans. Pattern Anal. Mach. Intell..

[33]  Leo Breiman,et al.  Bagging Predictors , 1996, Machine Learning.

[34]  Geoff Holmes,et al.  Classifier chains for multi-label classification , 2009, Machine Learning.

[35]  Tao Li,et al.  A comparative study on content-based music genre classification , 2003, SIGIR.