Quantum clustering-based weighted linear programming support vector regression for multivariable nonlinear problem

Linear programming support vector regression shows improved reliability and generates sparse solution, compared with standard support vector regression. We present the v-linear programming support vector regression approach based on quantum clustering and weighted strategy to solve the multivariable nonlinear regression problem. First, the method applied quantum clustering to variable selection, introduced inertia weight, and took prediction precision of v-linear programming support vector regression as evaluation criteria, which effectively removed redundancy feature attributes and also reduced prediction error and support vectors. Second, it proposed a new weighted strategy due to each data point having different influence on regression model and determined the weighted parameter p in terms of distribution of training error, which greatly improved the generalization approximate ability. Experimental results demonstrated that the proposed algorithm enabled the mean squared error of test sets of Boston housing, Bodyfat, Santa dataset to, respectively, decrease by 23.18, 78.52, and 41.39%, and also made support vectors degrade rapidly, relative to the original v-linear programming support vector regression method. In contrast with other methods exhibited in the relevant literatures, the present algorithm achieved better generalization performance.

[1]  Isabelle Guyon,et al.  An Introduction to Variable and Feature Selection , 2003, J. Mach. Learn. Res..

[2]  Assaf Gottlieb,et al.  Algorithm for data clustering in pattern recognition problems based on quantum mechanics. , 2001, Physical review letters.

[3]  Johan A. K. Suykens,et al.  A support vector machine formulation to PCA analysis and its kernel version , 2003, IEEE Trans. Neural Networks.

[4]  Gilles Brassard,et al.  Quantum clustering algorithms , 2007, ICML '07.

[5]  Alexander J. Smola,et al.  Support Vector Method for Function Approximation, Regression Estimation and Signal Processing , 1996, NIPS.

[6]  Alexander J. Smola,et al.  Support Vector Regression Machines , 1996, NIPS.

[7]  Lijuan Cao,et al.  A comparison of PCA, KPCA and ICA for dimensionality reduction in support vector machine , 2003, Neurocomputing.

[8]  Bernhard Schölkopf,et al.  A tutorial on support vector regression , 2004, Stat. Comput..

[9]  Li Zhang,et al.  Linear programming support vector machines , 2002, Pattern Recognit..

[10]  A. Asuncion,et al.  UCI Machine Learning Repository, University of California, Irvine, School of Information and Computer Sciences , 2007 .

[11]  Alexander J. Smola,et al.  Learning with kernels , 1998 .

[12]  Noboru Murata,et al.  Support Vector Machines for Linear Programming: Motivation and Formulations , 1999 .

[13]  I K Fodor,et al.  A Survey of Dimension Reduction Techniques , 2002 .

[14]  Wenming Zheng,et al.  An Improved Algorithm for Kernel Principal Component Analysis , 2005, Neural Processing Letters.

[15]  Erkki Oja,et al.  Independent component analysis: algorithms and applications , 2000, Neural Networks.

[16]  Bernhard Schölkopf,et al.  Kernel Principal Component Analysis , 1997, ICANN.

[17]  Dan A. Simovici,et al.  On feature selection through clustering , 2005, Fifth IEEE International Conference on Data Mining (ICDM'05).

[18]  David Horn,et al.  The Method of Quantum Clustering , 2001, NIPS.

[19]  David R. Musicant,et al.  Large Scale Kernel Regression via Linear Programming , 2002, Machine Learning.

[20]  B. Schölkopf,et al.  Linear programs for automatic accuracy control in regression. , 1999 .

[21]  Vladimir N. Vapnik,et al.  The Nature of Statistical Learning Theory , 2000, Statistics for Engineering and Information Science.

[22]  Vladimir Vapnik,et al.  Statistical learning theory , 1998 .

[23]  Beatrice Lazzerini,et al.  A new fuzzy relational clustering algorithm based on the fuzzy C-means algorithm , 2005, Soft Comput..

[24]  Carla E. Brodley,et al.  Feature Selection for Unsupervised Learning , 2004, J. Mach. Learn. Res..

[25]  Chih-Jen Lin,et al.  Combining SVMs with Various Feature Selection Strategies , 2006, Feature Extraction.

[26]  Johan A. K. Suykens,et al.  Weighted least squares support vector machines: robustness and sparse approximation , 2002, Neurocomputing.

[27]  Sushant Sachdeva,et al.  Dimension Reduction , 2008, Encyclopedia of GIS.

[28]  I. Jolliffe Principal Component Analysis , 2002 .

[29]  Bernhard Schölkopf,et al.  Shrinking the Tube: A New Support Vector Regression Algorithm , 1998, NIPS.