Kronecker Factorization for Speeding up Kernel Machines

In kernel machines, such as kernel principal component analysis (KPCA), Gaussian Processes (GPs), and Support Vector Machines (SVMs), the computational complexity of finding a solution is O(n), where n is the number of training instances. To reduce this expensive computational complexity, we propose using Kronecker factorization, which approximates a positive definite kernel matrix by the Kronecker product of two smaller positive definite matrices. This approximation can speed up the calculation of the kernel-matrix inverse or eigendecomposition involved in kernel machines. When the two factorized matrices have about the same dimensions, the computational complexity is improved from O(n) to O(n). We propose two methods to carry out Kronecker factorization and apply them to speed up KPCA. In Experiments show that our methods can drastically reduce the computation time of kernel machines without any significant degradation in their effectiveness.

[1]  R. Barham,et al.  An Algorithm for Least Squares Estimation of Nonlinear Parameters When Some of the Parameters Are Linear , 1972 .

[2]  Charles R. Johnson,et al.  Topics in Matrix Analysis , 1991 .

[3]  C. Loan,et al.  Approximation with Kronecker Products , 1992 .

[4]  Federico Girosi,et al.  Reducing the run-time complexity of Support Vector Machines , 1999 .

[5]  Bernhard Schölkopf,et al.  Nonlinear Component Analysis as a Kernel Eigenvalue Problem , 1998, Neural Computation.

[6]  J. Platt Sequential Minimal Optimization : A Fast Algorithm for Training Support Vector Machines , 1998 .

[7]  Catherine Blake,et al.  UCI Repository of machine learning databases , 1998 .

[8]  David Barber,et al.  Bayesian Classification With Gaussian Processes , 1998, IEEE Trans. Pattern Anal. Mach. Intell..

[9]  Bernhard Schölkopf,et al.  Sparse Greedy Matrix Approximation for Machine Learning , 2000, International Conference on Machine Learning.

[10]  Christopher K. I. Williams,et al.  Using the Nyström Method to Speed Up Kernel Machines , 2000, NIPS.

[11]  Bernhard Schölkopf,et al.  Sampling Techniques for Kernel Methods , 2001, NIPS.

[12]  E. D. Klerk,et al.  Aspects of semidefinite programming : interior point algorithms and selected applications , 2002 .

[13]  Kiyoshi Asai,et al.  The em Algorithm for Kernel Matrix Completion with Auxiliary Data , 2003, J. Mach. Learn. Res..

[14]  Zhihua Zhang,et al.  Surrogate maximization/minimization algorithms for AdaBoost and the logistic regression model , 2004, ICML '04.