Support vector machine training using matrix completion techniques

We combine interior-point methods and results from matrix completion theory in an approximate method for the large dense quadratic programming problems that arise in support vector machine training. The basic idea is to replace the dense kernel matrix with the maximum determinant positive definite completion of a subset of the entries of the kernel matrix. The resulting approximate kernel matrix has a sparse inverse and this property can be exploited to dramatically improve the efficiency of interior-point methods. If the sparsity pattern is chordal, the sparse inverse and its Cholesky factors are easily computed by efficient recursive algorithms. Numerical experiments with training sets of size up to 60000 show that the approximate interiorpoint approach is competitive with the libsvm package in terms of error rate and speed. The completion technique can also be applied to other dense convex optimization problems arising in machine learning, for example, in Gaussian process classification.

[1]  Chih-Jen Lin,et al.  LIBSVM: A library for support vector machines , 2011, TIST.

[2]  Evgueni A. Haroutunian,et al.  Information Theory and Statistics , 2011, International Encyclopedia of Statistical Science.

[3]  Robert H. Halstead,et al.  Matrix Computations , 2011, Encyclopedia of Parallel Computing.

[4]  Thorsten Joachims,et al.  Sparse kernel SVMs via cutting-plane training , 2009, Machine Learning.

[5]  Carl E. Rasmussen,et al.  Gaussian processes for machine learning , 2005, Adaptive computation and machine learning.

[6]  Monique Laurent,et al.  Matrix Completion Problems , 2009, Encyclopedia of Optimization.

[7]  Vwani P. Roychowdhury,et al.  Covariance selection for nonchordal graphs via chordal embedding , 2008, Optim. Methods Softw..

[8]  Chih-Jen Lin,et al.  A dual coordinate descent method for large-scale linear SVM , 2008, ICML '08.

[9]  Chih-Jen Lin,et al.  LIBLINEAR: A Library for Large Linear Classification , 2008, J. Mach. Learn. Res..

[10]  Thorsten Joachims,et al.  Training linear SVMs in linear time , 2006, KDD '06.

[11]  Yann LeCun,et al.  The mnist database of handwritten digits , 2005 .

[12]  Michael C. Ferris,et al.  Interior-Point Methods for Massive Support Vector Machines , 2002, SIAM J. Optim..

[13]  Katya Scheinberg,et al.  Efficient SVM Training Using Low-Rank Kernel Representations , 2002, J. Mach. Learn. Res..

[14]  D K Smith,et al.  Numerical Optimization , 2001, J. Oper. Res. Soc..

[15]  Bernhard Schölkopf,et al.  Sparse Greedy Matrix Approximation for Machine Learning , 2000, International Conference on Machine Learning.

[16]  John C. Platt,et al.  Fast training of support vector machines using sequential minimal optimization, advances in kernel methods , 1999 .

[17]  Bernhard Schölkopf,et al.  Nonlinear Component Analysis as a Kernel Eigenvalue Problem , 1998, Neural Computation.

[18]  Alexander J. Smola,et al.  Learning with kernels , 1998 .

[19]  Thorsten Joachims,et al.  Making large scale SVM learning practical , 1998 .

[20]  Federico Girosi,et al.  An improved training algorithm for support vector machines , 1997, Neural Networks for Signal Processing VII. Proceedings of the 1997 IEEE Signal Processing Society Workshop.

[21]  Stephen J. Wright Primal-Dual Interior-Point Methods , 1997, Other Titles in Applied Mathematics.

[22]  Thomas M. Cover,et al.  Elements of Information Theory , 2005 .

[23]  Charles R. Johnson,et al.  Positive definite completions of partial Hermitian matrices , 1984 .