An Approach of Feature Subset Selection Using Simulated Quantum Annealing

Feature selection is one of the important preprocessing steps in machine learning and data mining domain. However, finding the best feature subsets for large datasets is computationally expensive task. Meanwhile, quantum computing has emerged as a new computational model that is able to speed up many classical computationally expensive problems. Annealing-based quantum model, for example, finds the lowest energy state of an Ising model Hamiltonian, which is the formalism for Quadratic Unconstrained Binary Optimization (QUBO). Due to its capabilities in producing quality solution to the hard combinatorial optimization problems with less computational effort, quantum annealing has the potentiality in feature subset selection. Although several hard optimization problems are solved, using quantum annealing, not sufficient work has been done on quantum annealing based feature subset selection. Though the reported approaches have good theoretical foundation, they usually lack required empirical rigor. In this paper, we attempt to reduce classical benchmark feature evaluation functions like mRMR, JMI, and FCBF to QUBO formulation, enabling the use of quantum annealing based optimization to feature selection. We then apply QUBO of ten datasets using both Simulated Annealing (SA) and Simulated Quantum Annealing (SQA) and compared the result. Our empirical results confirm that, for seven in ten datasets, SQA is able to produce at most equal or less number of features in all selected subset compared to SA does. SQA also results in stable feature subsets for all datasets.

[1]  Huan Liu,et al.  Feature Selection for High-Dimensional Data: A Fast Correlation-Based Filter Solution , 2003, ICML.

[2]  Phil Goddard,et al.  Optimal Feature Selection Using a Quantum Annealer , 2018 .

[3]  Basabi Chakraborty,et al.  An empirical study of feature selection for classification using genetic algorithm , 2018 .

[4]  H. Nishimori,et al.  Quantum annealing in the transverse Ising model , 1998, cond-mat/9804280.

[5]  Kristan Temme,et al.  Supervised learning with quantum-enhanced feature spaces , 2018, Nature.

[6]  Kathryn A. Dowsland,et al.  Simulated Annealing , 1989, Encyclopedia of GIS.

[7]  E. Ising Beitrag zur Theorie des Ferromagnetismus , 1925 .

[8]  Aram Wettroth Harrow,et al.  Simulated Quantum Annealing Can Be Exponentially Faster Than Classical Simulated Annealing , 2016, 2016 IEEE 57th Annual Symposium on Foundations of Computer Science (FOCS).

[9]  I. Chuang,et al.  Quantum Computation and Quantum Information: Bibliography , 2010 .

[10]  Kewei Cheng,et al.  Feature Selection , 2016, ACM Comput. Surv..

[11]  Jacob biamonte,et al.  Quantum machine learning , 2016, Nature.

[12]  Gavin Brown,et al.  Conditional Likelihood Maximisation: A Unifying Framework for Information Theoretic Feature Selection , 2012, J. Mach. Learn. Res..

[13]  Frank Mueller,et al.  Programming quantum computers: a primer with IBM Q and D-Wave exercises , 2019, PPoPP.

[14]  S. Sitharama Iyengar,et al.  Data-Driven Techniques in Disaster Information Management , 2017, ACM Comput. Surv..

[15]  Simone Severini,et al.  Quantum machine learning: a classical perspective , 2017, Proceedings of the Royal Society A: Mathematical, Physical and Engineering Sciences.

[16]  David Von Dollen,et al.  Traffic Flow Optimization Using a Quantum Annealer , 2017, Front. ICT.

[17]  Huan Liu,et al.  Feature selection for classification: A review , 2014 .

[18]  Biao Wu,et al.  Exact Equivalence between Quantum Adiabatic Algorithm and Quantum Circuit Algorithm , 2017, Chinese Physics Letters.

[19]  Christian F. A. Negre,et al.  Graph Partitioning using Quantum Annealing on the D-Wave System , 2017, ArXiv.

[20]  Zhiming Huang,et al.  Quantum-enhanced feature selection with forward selection and backward elimination , 2018, Quantum Inf. Process..

[21]  Fuhui Long,et al.  Feature selection based on mutual information criteria of max-dependency, max-relevance, and min-redundancy , 2003, IEEE Transactions on Pattern Analysis and Machine Intelligence.