On a new approach for Lagrangian support vector regression

In this paper, a simplification of the necessary and sufficient Karush–Kuhn–Tucker (KKT) optimality conditions for the Lagrangian support vector regression in 2-norm resulting a naïve root finding problem of a system of equations in m variables is proposed, where m is the number of training examples. However, since the resulting system of equations contains terms having the nonsmooth “plus” function, two approaches are considered: (i) the problem is reformulated into an equivalent absolute value equation problem and solved by functional iterative and generalized Newton methods and (ii) solve the original root finding problem by generalized Newton method. The proofs of convergence and pseudo-codes of the iterative methods are given. Numerical experiments performed on a number of synthetic and real-world benchmark datasets showing similar generalization performance with much faster learning speed in comparison with support vector regression (SVR) and as good as performance in comparison with the least squares SVR, unconstrained Lagrangian SVR proposed in Balasundaram et al. (Neural Netw 51:67–79, 2014), twin SVR and extreme learning machine clearly indicate the effectiveness and suitability of the proposed problem formulation solved by functional iterative and generalized Newton methods.

[1]  Xinjun Peng,et al.  TSVR: An efficient Twin Support Vector Machine for regression , 2010, Neural Networks.

[2]  Lennart Ljung,et al.  Nonlinear black-box modeling in system identification: a unified overview , 1995, Autom..

[3]  Jason Weston,et al.  Gene Selection for Cancer Classification using Support Vector Machines , 2002, Machine Learning.

[4]  Shifei Ding,et al.  An overview on twin support vector machines , 2012, Artificial Intelligence Review.

[5]  S. Balasundaram,et al.  Training Lagrangian twin support vector regression via unconstrained convex minimization , 2014, Knowl. Based Syst..

[6]  Nello Cristianini,et al.  An Introduction to Support Vector Machines and Other Kernel-based Learning Methods , 2000 .

[7]  Olvi L. Mangasarian,et al.  A generalized Newton method for absolute value equations , 2009, Optim. Lett..

[8]  Chee Kheong Siew,et al.  Extreme learning machine: Theory and applications , 2006, Neurocomputing.

[9]  S. Balasundaram,et al.  Lagrangian support vector regression via unconstrained convex minimization , 2014, Neural Networks.

[10]  Thorsten Joachims,et al.  Text Categorization with Support Vector Machines: Learning with Many Relevant Features , 1998, ECML.

[11]  Guilherme A. Barreto,et al.  Nonlinear System Identification Using Local ARX Models Based On The Self-Organazing Map , 2006 .

[12]  Hongming Zhou,et al.  Extreme Learning Machine for Regression and Multiclass Classification , 2012, IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics).

[13]  Gene H. Golub,et al.  Matrix computations (3rd ed.) , 1996 .

[14]  S. Balasundaram,et al.  Finite Newton method for implicit Lagrangian support vector regression , 2011, Int. J. Knowl. Based Intell. Eng. Syst..

[15]  David R. Musicant,et al.  Lagrangian Support Vector Machines , 2001, J. Mach. Learn. Res..

[16]  Catherine Blake,et al.  UCI Repository of machine learning databases , 1998 .

[17]  Min Wang,et al.  Seeking multi-thresholds directly from support vectors for image segmentation , 2005, Neurocomputing.

[18]  A. Gretton,et al.  Support vector regression for black-box system identification , 2001, Proceedings of the 11th IEEE Signal Processing Workshop on Statistical Signal Processing (Cat. No.01TH8563).

[19]  David R. Musicant,et al.  Active Support Vector Machine Classification , 2000, NIPS.

[20]  Gene H. Golub,et al.  Matrix computations , 1983 .

[21]  Zhongzhi Shi,et al.  Primal least squares twin support vector regression , 2013, Journal of Zhejiang University SCIENCE C.

[22]  Johan A. K. Suykens,et al.  Least Squares Support Vector Machine Classifiers , 1999, Neural Processing Letters.

[23]  Reshma Khemchandani,et al.  Twin Support Vector Machines for Pattern Classification , 2007, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[24]  Janez Demsar,et al.  Statistical Comparisons of Classifiers over Multiple Data Sets , 2006, J. Mach. Learn. Res..

[25]  S. Balasundaram,et al.  On Lagrangian support vector regression , 2010, Expert Syst. Appl..

[26]  Guilherme A. Barreto,et al.  NONLINEAR SYSTEM IDENTIFICATION USING LOCAL ARX MODELS BASED ON THE SELF-ORGANIZING MAP , 2008 .

[27]  Vladimir N. Vapnik,et al.  The Nature of Statistical Learning Theory , 2000, Statistics for Engineering and Information Science.

[28]  Federico Girosi,et al.  Training support vector machines: an application to face detection , 1997, Proceedings of IEEE Computer Society Conference on Computer Vision and Pattern Recognition.

[29]  Glenn Fung,et al.  A Feature Selection Newton Method for Support Vector Machine Classification , 2004, Comput. Optim. Appl..

[30]  David R. Musicant,et al.  Active set support vector regression , 2004, IEEE Transactions on Neural Networks.