An optimal feature selection technique using the concept of mutual information

We present a mutual information-based technique to perform feature selection for the purpose of classification. The technique selects those features that have maximum mutual information with the specified classes. The best solution may be obtained through an exhaustive search (all possible combinations). However, even with a small number of features, this solution becomes impractical due to the exponentially increasing computational cost. Unlike other techniques that select features individually, our technique considers a trade off between computational cost and combined feature selection. Extensive experiments have shown that the proposed technique outperforms existing feature selection methods based on individual features.

[1]  Jim R. Parker,et al.  Algorithms for image processing and computer vision , 1996 .

[2]  Pierre Comon,et al.  Independent component analysis, A new concept? , 1994, Signal Process..

[3]  Erkki Oja,et al.  Neural Networks, Principal Components, and Subspaces , 1989, Int. J. Neural Syst..

[4]  Fraser,et al.  Independent coordinates for strange attractors from mutual information. , 1986, Physical review. A, General physics.

[5]  Mohamed A. Deriche,et al.  A hybrid information maximisation (HIM) algorithm for optimal feature selection from multi-channel data , 2000, 2000 IEEE International Conference on Acoustics, Speech, and Signal Processing. Proceedings (Cat. No.00CH37100).

[6]  P. Comon Independent Component Analysis , 1992 .

[7]  Roberto Battiti,et al.  Using mutual information for selecting features in supervised neural net learning , 1994, IEEE Trans. Neural Networks.

[8]  E. Oja,et al.  Independent Component Analysis , 2013 .