A Fast Feature Selection Algorithm Based on Swarm Intelligence in Acoustic Defect Detection

Feature selection plays an important role in data mining and pattern recognition. However, most existing feature selection methods suffer from stagnation in local optimal and/or high computational cost. In general, the feature selection process can be considered as a global optimization problem. The swarm intelligence algorithm can effectively find the optimal feature subset due to its global search ability. But it usually consumes very long running time when dealing with large data sets. In this paper, feature selection is transformed into a global optimization problem, which provides a fast and efficient method based on swarm intelligence algorithm. First, we propose a global optimization framework for filter-based feature selection and its mathematical model. Furthermore, to solve the feature selection problem for acoustic defect detection, we combine the shuffled frog leaping algorithm (SFLA) with an improved minimum-redundancy maximum-relevancy (ImRMR), named SFLA-ImRMR. In the experiments, a back propagation neural network is employed to evaluate the classification performance of the selected feature subset on the test sets of acoustic defect detection. The results show that SFLA-ImRMR achieved similar performance to the other algorithms within the shortest time.

[1]  Thomas Marill,et al.  On the effectiveness of receptors in recognition systems , 1963, IEEE Trans. Inf. Theory.

[2]  Myong Kee Jeong,et al.  An evolutionary algorithm with the partial sequential forward floating search mutation for large-scale feature selection problems , 2015, J. Oper. Res. Soc..

[3]  Pablo A. Estévez,et al.  A review of feature selection methods based on mutual information , 2013, Neural Computing and Applications.

[4]  Wang Liping,et al.  Feature Selection Algorithm Based on Conditional Dynamic Mutual Information , 2015 .

[5]  Fuhui Long,et al.  Feature selection based on mutual information criteria of max-dependency, max-relevance, and min-redundancy , 2003, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[6]  Ivor W. Tsang,et al.  The Emerging "Big Dimensionality" , 2014, IEEE Computational Intelligence Magazine.

[7]  P. Langley Selection of Relevant Features in Machine Learning , 1994 .

[8]  A. Wayne Whitney,et al.  A Direct Method of Nonparametric Measurement Selection , 1971, IEEE Transactions on Computers.

[9]  Mengjie Zhang,et al.  Particle Swarm Optimization for Feature Selection in Classification: A Multi-Objective Approach , 2013, IEEE Transactions on Cybernetics.

[10]  Huan Liu,et al.  Efficient Feature Selection via Analysis of Relevance and Redundancy , 2004, J. Mach. Learn. Res..

[11]  Zehang Sun,et al.  Object detection using feature subset selection , 2004, Pattern Recognit..

[12]  Xin Yao,et al.  A Survey on Evolutionary Computation Approaches to Feature Selection , 2016, IEEE Transactions on Evolutionary Computation.

[13]  Huan Liu,et al.  Feature Selection for Classification , 1997, Intell. Data Anal..

[14]  Mengjie Zhang,et al.  Binary particle swarm optimisation for feature selection: A filter based approach , 2012, 2012 IEEE Congress on Evolutionary Computation.

[15]  Jacek M. Zurada,et al.  Normalized Mutual Information Feature Selection , 2009, IEEE Transactions on Neural Networks.

[16]  Anil K. Jain,et al.  39 Dimensionality and sample size considerations in pattern recognition practice , 1982, Classification, Pattern Recognition and Reduction of Dimensionality.

[17]  Keinosuke Fukunaga,et al.  A Branch and Bound Algorithm for Feature Subset Selection , 1977, IEEE Transactions on Computers.

[18]  Huan Liu,et al.  Feature Selection: An Ever Evolving Frontier in Data Mining , 2010, FSDM.

[19]  Magdalene Marinaki,et al.  A bumble bees mating optimization algorithm for the feature selection problem , 2016, Int. J. Mach. Learn. Cybern..

[20]  Zengyou He,et al.  Stable Feature Selection for Biomarker Discovery , 2010, Comput. Biol. Chem..

[21]  Haibin Duan,et al.  New progresses in swarm intelligence-based computation , 2015, Int. J. Bio Inspired Comput..

[22]  Huan Liu,et al.  Toward integrating feature selection algorithms for classification and clustering , 2005, IEEE Transactions on Knowledge and Data Engineering.

[23]  Alireza Rahimi-Vahed,et al.  A hybrid multi-objective shuffled frog-leaping algorithm for a mixed-model assembly line sequencing problem , 2007, Comput. Ind. Eng..

[24]  Kevin E Lansey,et al.  Optimization of Water Distribution Network Design Using the Shuffled Frog Leaping Algorithm , 2003 .

[25]  Huan Liu,et al.  Manipulating Data and Dimension Reduction Methods: Feature Selection , 2009, Encyclopedia of Complexity and Systems Science.

[26]  Daoliang Li,et al.  Feature selection based on improved ant colony optimization for online detection of foreign fiber in cotton , 2014, Appl. Soft Comput..

[27]  Jennie Si,et al.  FREL: A Stable Feature Selection Algorithm , 2015, IEEE Transactions on Neural Networks and Learning Systems.

[28]  Leslie S. Smith,et al.  Feature subset selection in large dimensionality domains , 2010, Pattern Recognit..

[29]  Anne M. P. Canuto,et al.  A comparative analysis of genetic algorithm and ant colony optimization to select attributes for an heterogeneous ensemble of classifiers , 2010, IEEE Congress on Evolutionary Computation.