Feature selection with limited bit depth mutual information for portable embedded systems

Abstract Since wearable computing systems have grown in importance in the last years, there is an increased interest in implementing machine learning algorithms with reduced precision parameters/computations. Not only learning, also feature selection, most of the times a mandatory preprocessing step in machine learning, is often constrained by the available computational resources. This work considers mutual information – one of the most common measures of dependence used in feature selection algorithms – with a limited number of bits. In order to test the procedure designed, we have implemented it in several well-known feature selection algorithms. Experimental results over several synthetic and real datasets demonstrate that low bit representations are sufficient to achieve performances close to that of double precision parameters and thus open the door for the use of feature selection in embedded platforms that minimize the energy consumption and carbon emissions.

[1]  D. E. Roberts,et al.  The Upper Tail Probabilities of Spearman's Rho , 1975 .

[2]  Nazar Khan,et al.  Machine Learning at the Network Edge: A Survey , 2019, ACM Comput. Surv..

[3]  Verónica Bolón-Canedo,et al.  Recent advances and emerging challenges of feature selection in the context of big data , 2015, Knowl. Based Syst..

[4]  Gavin Brown,et al.  Efficient feature selection using shrinkage estimators , 2019, Machine Learning.

[5]  Wei-Yin Loh,et al.  Classification and regression trees , 2011, WIREs Data Mining Knowl. Discov..

[6]  Fuhui Long,et al.  Feature selection based on mutual information criteria of max-dependency, max-relevance, and min-redundancy , 2003, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[7]  Gavin Brown,et al.  Conditional Likelihood Maximisation: A Unifying Framework for Information Theoretic Feature Selection , 2012, J. Mach. Learn. Res..

[8]  M. Tesmer,et al.  AMIFS: adaptive feature selection by using mutual information , 2004, 2004 IEEE International Joint Conference on Neural Networks (IEEE Cat. No.04CH37541).

[9]  David D. Lewis,et al.  Feature Selection and Feature Extraction for Text Categorization , 1992, HLT.

[10]  Swarup Bhunia,et al.  Wearables, Implants, and Internet of Things: The Technology Needs in the Evolving Landscape , 2016, IEEE Transactions on Multi-Scale Computing Systems.

[11]  Mark S. Nixon,et al.  Gait Feature Subset Selection by Mutual Information , 2007, 2007 First IEEE International Conference on Biometrics: Theory, Applications, and Systems.

[12]  Ran El-Yaniv,et al.  Quantized Neural Networks: Training Neural Networks with Low Precision Weights and Activations , 2016, J. Mach. Learn. Res..

[13]  Gavin Brown,et al.  Simple strategies for semi-supervised feature selection , 2017, Machine Learning.

[14]  Roberto Battiti,et al.  Using mutual information for selecting features in supervised neural net learning , 1994, IEEE Trans. Neural Networks.

[15]  Sebastian Tschiatschek,et al.  Parameter Learning of Bayesian Network Classifiers Under Computational Constraints , 2015, ECML/PKDD.

[16]  Liam Paninski,et al.  Estimation of Entropy and Mutual Information , 2003, Neural Computation.

[17]  Christian Hennig,et al.  Recovering the number of clusters in data sets with noise features using feature rescaling factors , 2015, Inf. Sci..

[18]  John E. Moody,et al.  Data Visualization and Feature Selection: New Algorithms for Nongaussian Data , 1999, NIPS.

[19]  Philip Koopman Design Constraints on Embedded Real Time Control Systems , 1990 .

[20]  David W. Aha,et al.  Instance-Based Learning Algorithms , 1991, Machine Learning.

[21]  Song Han,et al.  EIE: Efficient Inference Engine on Compressed Deep Neural Network , 2016, 2016 ACM/IEEE 43rd Annual International Symposium on Computer Architecture (ISCA).

[22]  Janez Demsar,et al.  Statistical Comparisons of Classifiers over Multiple Data Sets , 2006, J. Mach. Learn. Res..

[23]  Samuel H. Huang Supervised feature selection: A tutorial , 2015, Artif. Intell. Res..

[24]  Verónica Bolón-Canedo,et al.  Feature Selection with Limited Bit Depth Mutual Information for Embedded Systems , 2018 .

[25]  Pedro Larrañaga,et al.  A review of feature selection techniques in bioinformatics , 2007, Bioinform..

[26]  Juan José Rodríguez Diez,et al.  On feature selection protocols for very low-sample-size data , 2018, Pattern Recognit..