Sparse Least Squares Support Vector Regressors Trained in the Reduced Empirical Feature Space

In this paper we discuss sparse least squares support vector regressors (sparse LS SVRs) defined in the reduced empirical feature space, which is a subspace of mapped training data. Namely, we define an LS SVR in the primal form in the empirical feature space, which results in solving a set of linear equations. The independent components in the empirical feature space are obtained by deleting dependent components during the Cholesky factorization of the kernel matrix. The independent components are associated with support vectors and controlling the threshold of the Cholesky factorization we obtain a sparse LS SVM. For linear kernels the number of support vectors is the number of input variables at most and if we use the input axes as support vectors, the primal and dual forms are equivalent. By computer experiments we show that we can reduce the number of support vectors without deteriorating the generalization ability.

[1]  Christopher J. C. Burges,et al.  Simplified Support Vector Decision Rules , 1996, ICML.

[2]  S. Sathiya Keerthi,et al.  Building Support Vector Machines with Reduced Classifier Complexity , 2006, J. Mach. Learn. Res..

[3]  Weidong Zhang,et al.  Improved sparse least-squares support vector machine classifiers , 2006, Neurocomputing.

[4]  Shigeo Abe,et al.  Sparse least squares support vector training in the reduced empirical feature space , 2007, Pattern Analysis and Applications.

[5]  Shigeo Abe,et al.  Support Vector Machines for Pattern Classification (Advances in Pattern Recognition) , 2005 .

[6]  Enric Hernández,et al.  Indistinguishability relations in Dempster-Shafer theory of evidence , 2004, Int. J. Approx. Reason..

[7]  Kenji Baba,et al.  Explicit representation of knowledge acquired from plant historical data using neural network , 1990, 1990 IJCNN International Joint Conference on Neural Networks.

[8]  Bernhard Schölkopf,et al.  A Direct Method for Building Sparse Kernel Learning Algorithms , 2006, J. Mach. Learn. Res..

[9]  M. Omair Ahmad,et al.  Optimizing the kernel in the empirical feature space , 2005, IEEE Transactions on Neural Networks.

[10]  D. Rubinfeld,et al.  Hedonic housing prices and the demand for clean air , 1978 .

[11]  Johan A. K. Suykens,et al.  Least Squares Support Vector Machine Classifiers , 1999, Neural Processing Letters.

[12]  R. Scott Crowder,et al.  Predicting the Mackey-Glass Timeseries With Cascade-Correlation Learning , 1990 .

[13]  Johan A. K. Suykens,et al.  Least Squares Support Vector Machines , 2002 .

[14]  Liu Xiao,et al.  A Sparse Least Squares Support Vector Machine Classifier , 2007 .

[15]  Shigeo Abe,et al.  KPCA-based training of a kernel fuzzy classifier with ellipsoidal regions , 2004, Int. J. Approx. Reason..