Double-fold localized multiple matrixized learning machine

In this paper, we develop an effective multiple-matrixized learning machine named Double-fold Localized Multiple Matrixized Learning Machine (DLMMLM). The characteristic of the proposed DLMMLM is that it possesses double folds of local information from data. The first fold lies in the whole representation space which consists of different matrix representations. It is known that each pattern can be represented by different matrix representations. The matrices have their respective representation information and can play different discriminant roles in the final classification. Therefore from the viewpoint of the whole representation space, each matrix has its own local information. The second fold is that in each matrix representation learning, different patterns represented with the same matrix representation can carry different information. Therefore in the pattern space with the same matrix size, local information of different patterns should be introduced into the classifier design. On the whole, the advantages of the proposed DLMMLM are: (i) establishing a pattern-depended function in the matrixized learning so as to realize different roles of patterns for the first time; (ii) adopting the double-fold local information in both the representation space and the pattern space; (iii) proposing a new nonlinear classifier that is different from the state-of-the-art kernelization one; and (iv) getting a tighter empirical generalization risk bound in terms of the Rademacher complexity and thus achieving a statistically superior classification performance than those classifiers without the introduction of the double-fold local information.

[1]  Robert Sabourin,et al.  “One Against One” or “One Against All”: Which One is Better for Handwriting Recognition with SVMs? , 2006 .

[2]  Guiqiang Ni,et al.  One-Class Support Vector Machines Based on Matrix Patterns , 2011 .

[3]  Tomaso Poggio,et al.  Image Representations for Visual Learning , 1996, Science.

[4]  Ryan M. Rifkin,et al.  In Defense of One-Vs-All Classification , 2004, J. Mach. Learn. Res..

[5]  William Stafford Noble,et al.  Support vector machine , 2013 .

[6]  Douglas C. Montgomery,et al.  Response Surface Methodology: Process and Product Optimization Using Designed Experiments , 1995 .

[7]  Vladimir Vapnik,et al.  Chervonenkis: On the uniform convergence of relative frequencies of events to their probabilities , 1971 .

[8]  Xuelong Li,et al.  Single-image super-resolution via local learning , 2011, Int. J. Mach. Learn. Cybern..

[9]  Jing Peng,et al.  SVM vs regularized least squares classification , 2004, ICPR 2004.

[10]  Rameswar Debnath,et al.  A decision based one-against-one method for multi-class support vector machine , 2004, Pattern Analysis and Applications.

[12]  Patrick P. K. Chan,et al.  Dynamic fusion method using Localized Generalization Error Model , 2012, Inf. Sci..

[13]  Jin Xu,et al.  Multiple empirical kernel learning based on local information , 2012, Neural Computing and Applications.

[14]  Takashi Takenouchi,et al.  Statistical Learning Theory by Boosting Method , 2004 .

[15]  Catherine Blake,et al.  UCI Repository of machine learning databases , 1998 .

[16]  Ethem Alpaydin Multiple Kernel Machines Using Localized Kernels , 2009 .

[17]  Jin Xu,et al.  Regularized multi-view learning machine based on response surface technique , 2012, Neurocomputing.

[18]  Sayan Mukherjee,et al.  Choosing Multiple Parameters for Support Vector Machines , 2002, Machine Learning.

[19]  Songcan Chen,et al.  Matrix-pattern-oriented Ho-Kashyap classifier with regularization learning , 2007, Pattern Recognit..

[20]  Shan Suthaharan,et al.  Support Vector Machine , 2016 .

[21]  Patrick P. K. Chan,et al.  Radial Basis Function network learning using localized generalization error bound , 2009, Inf. Sci..

[22]  Ethem Alpayd Multiple Kernel Machines Using Localized Kernels , 2009 .

[23]  Shigeo Abe DrEng Pattern Classification , 2001, Springer London.

[24]  Shahar Mendelson,et al.  Rademacher averages and phase transitions in Glivenko-Cantelli classes , 2002, IEEE Trans. Inf. Theory.

[25]  Songcan Chen,et al.  New Least Squares Support Vector Machines Based on Matrix Patterns , 2007, Neural Processing Letters.

[26]  R. C. Williamson,et al.  Generalization Bounds via Eigenvalues of the Gram matrix , 1999 .

[27]  V. Koltchinskii,et al.  Rademacher Processes and Bounding the Risk of Function Learning , 2004, math/0405338.

[28]  Songcan Chen,et al.  Matrix-pattern-oriented least squares support vector classifier with AdaBoost , 2008, Pattern Recognit. Lett..

[29]  Jieping Ye,et al.  Generalized Low Rank Approximations of Matrices , 2004, Machine Learning.

[30]  Jie Ding,et al.  Optimization of culture conditions for hydrogen production by Ethanoligenens harbinense B49 using response surface methodology. , 2009, Bioresource technology.

[31]  Corinna Cortes,et al.  Support-Vector Networks , 1995, Machine Learning.

[32]  Yi-Zeng Liang,et al.  Monte Carlo cross validation , 2001 .

[33]  Narendra Ahuja,et al.  Rank-R approximation of tensors using image-as-matrix representation , 2005, 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05).

[34]  Michael I. Jordan,et al.  Multiple kernel learning, conic duality, and the SMO algorithm , 2004, ICML.

[35]  Daoqiang Zhang,et al.  Feature extraction approaches based on matrix pattern: MatPCA and MatFLDA , 2005, Pattern Recognit. Lett..

[36]  Janez Demsar,et al.  Statistical Comparisons of Classifiers over Multiple Data Sets , 2006, J. Mach. Learn. Res..

[37]  Ethem Alpaydin,et al.  Localized Multiple Kernel Regression , 2010, 2010 20th International Conference on Pattern Recognition.

[38]  A. Zinatizadeh,et al.  Application of response surface methodology (RSM) to optimize coagulation-flocculation treatment of leachate using poly-aluminum chloride (PAC) and alum. , 2009, Journal of hazardous materials.

[39]  David G. Stork,et al.  Pattern Classification , 1973 .

[40]  Songcan Chen,et al.  A novel multi-view learning developed from single-view patterns , 2011, Pattern Recognit..

[41]  Tomaso A. Poggio,et al.  Regularization Networks and Support Vector Machines , 2000, Adv. Comput. Math..

[42]  Ethem Alpaydin,et al.  Localized multiple kernel learning , 2008, ICML '08.

[43]  Jacek M. Łȩski,et al.  Ho--Kashyap classifier with generalization control , 2003 .

[44]  Chih-Jen Lin,et al.  A comparison of methods for multiclass support vector machines , 2002, IEEE Trans. Neural Networks.

[45]  Ming Li,et al.  2D-LDA: A statistical linear discriminant analysis for image matrix , 2005, Pattern Recognit. Lett..

[46]  Alejandro F. Frangi,et al.  Two-dimensional PCA: a new approach to appearance-based face representation and recognition , 2004 .

[47]  Siuli Mukhopadhyay,et al.  Response surface methodology , 2010 .

[48]  Sankar K. Pal,et al.  Incorporating local image structure in normalized cut based graph partitioning for grouping of pixels , 2013, Inf. Sci..

[49]  Vladimir Koltchinskii,et al.  Rademacher penalties and structural risk minimization , 2001, IEEE Trans. Inf. Theory.

[50]  Peter L. Bartlett,et al.  Model Selection and Error Estimation , 2000, Machine Learning.

[51]  Zhe Wang,et al.  Three-fold structured classifier design based on matrix pattern , 2013, Pattern Recognit..