Multi-layer heterogeneous ensemble with classifier and feature selection

Deep Neural Networks have achieved many successes when applying to visual, text, and speech information in various domains. The crucial reasons behind these successes are the multi-layer architecture and the in-model feature transformation of deep learning models. These design principles have inspired other sub-fields of machine learning including ensemble learning. In recent years, there are some deep homogenous ensemble models introduced with a large number of classifiers in each layer. These models, thus, require a costly computational classification. Moreover, the existing deep ensemble models use all classifiers including unnecessary ones which can reduce the predictive accuracy of the ensemble. In this study, we propose a multi-layer ensemble learning framework called MUlti-Layer heterogeneous Ensemble System (MULES) to solve the classification problem. The proposed system works with a small number of heterogeneous classifiers to obtain ensemble diversity, therefore being efficiency in resource usage. We also propose an Evolutionary Algorithm-based selection method to select the subset of suitable classifiers and features at each layer to enhance the predictive performance of MULES. The selection method uses NSGA-II algorithm to optimize two objectives concerning classification accuracy and ensemble diversity. Experiments on 33 datasets confirm that MULES is better than a number of well-known benchmark algorithms.

[1]  Sung-Bae Cho,et al.  An Evolutionary Algorithm Approach to Optimal Ensemble Classifiers for DNA Microarray Data Analysis , 2008, IEEE Transactions on Evolutionary Computation.

[2]  Lev V. Utkin,et al.  A deep forest classifier with weights of class probability distribution subsets , 2019, Knowl. Based Syst..

[3]  TIEN THANH NGUYEN,et al.  Optimization of ensemble classifier system based on multiple objectives genetic algorithm , 2014, 2014 International Conference on Machine Learning and Cybernetics.

[4]  Janez Demsar,et al.  Statistical Comparisons of Classifiers over Multiple Data Sets , 2006, J. Mach. Learn. Res..

[5]  Mengjie Zhang,et al.  Evolving Deep Convolutional Neural Networks for Image Classification , 2017, IEEE Transactions on Evolutionary Computation.

[6]  ZhangPeng,et al.  When Ensemble Learning Meets Deep Learning , 2016 .

[7]  Witold Pedrycz,et al.  Combining heterogeneous classifiers via granular prototypes , 2018, Appl. Soft Comput..

[8]  Man Leung Wong,et al.  Optimizing stacking ensemble by an ant colony optimization approach , 2011, GECCO '11.

[9]  Hod Lipson,et al.  Autostacker: a compositional evolutionary learning system , 2018, GECCO.

[10]  Kalyanmoy Deb,et al.  A fast and elitist multiobjective genetic algorithm: NSGA-II , 2002, IEEE Trans. Evol. Comput..

[11]  Grigorios Tsoumakas,et al.  An ensemble uncertainty aware measure for directed hill climbing ensemble pruning , 2010, Machine Learning.

[12]  Jacob Schrum Evolving indirectly encoded convolutional neural networks to play tetris with low-level features , 2018, GECCO.

[13]  Xin Yao,et al.  A Survey on Evolutionary Computation Approaches to Feature Selection , 2016, IEEE Transactions on Evolutionary Computation.

[14]  Alan Wee-Chung Liew,et al.  Combining Multi Classifiers Based on a Genetic Algorithm - A Gaussian Mixture Model Framework , 2014, ICIC.

[15]  Jakub Nalepa,et al.  Memetic evolution of deep neural networks , 2018, GECCO.

[16]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[17]  Alan Wee-Chung Liew,et al.  Fusion of Classifiers Based on a Novel 2-Stage Model , 2014, ICMLC.

[18]  M. Bacauskienea,et al.  A feature selection technique for generation of classification committees and its application to categorization of laryngeal images , 2013 .

[19]  Yaochu Jin,et al.  Stacking-based ensemble learning of decision trees for interpretable prostate cancer detection , 2019, Appl. Soft Comput..

[20]  Alan L. Yuille,et al.  Genetic CNN , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[21]  Ji Feng,et al.  Deep Forest: Towards An Alternative to Deep Neural Networks , 2017, IJCAI.

[22]  Ludmila I. Kuncheva,et al.  Measures of Diversity in Classifier Ensembles and Their Relationship with the Ensemble Accuracy , 2003, Machine Learning.

[23]  Jane You,et al.  Progressive subspace ensemble learning , 2016, Pattern Recognit..

[24]  Tianqi Chen,et al.  XGBoost: A Scalable Tree Boosting System , 2016, KDD.

[25]  Paul A. Viola,et al.  Rapid object detection using a boosted cascade of simple features , 2001, Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition. CVPR 2001.

[26]  Guigang Zhang,et al.  Deep Learning , 2016, Int. J. Semantic Comput..

[27]  Leo Breiman,et al.  Random Forests , 2001, Machine Learning.

[28]  Alan Wee-Chung Liew,et al.  Simultaneous meta-data and meta-classifier selection in multiple classifier system , 2019, GECCO.

[29]  Jiri Matas,et al.  On Combining Classifiers , 1998, IEEE Trans. Pattern Anal. Mach. Intell..

[30]  Senén Barro,et al.  Do we need hundreds of classifiers to solve real world classification problems? , 2014, J. Mach. Learn. Res..

[31]  Bo Wang,et al.  When Ensemble Learning Meets Deep Learning: a New Deep Support Vector Machine for Classification , 2016, Knowl. Based Syst..