Reweighting Recognition Using Modified Kernel Principal Component Analysis via Manifold Learning

Kernel Principal Component Analysis (KPCA) is a widely used technique in the dimension reduction, de-noising and discovering nonlinear intrinsic dimensions of data set. In this paper we describe a reweighing kernel-based classification method for improving recognition problem. Firstly, we map the training samples to the feature space by non-linear transformation, and then perform principal component analysis(PCA) using the selected kernel function in the feature space, and get the linear representation of testing samples in the feature space. Secondly, by using the idea of reweighting, we select the similarity between testing sample and each training sample as the weight of reweighting, then take the final weight as the criteria of classification. The experimental results demonstrate that our method is more accurate than Support Vector Machine (SVM) classification method and Linear Discriminant Analysis (LDA) classification. In addition, the number of training samples that our method need is much smaller than some other methods.

[1]  Gunnar Rätsch,et al.  Kernel PCA and De-Noising in Feature Spaces , 1998, NIPS.

[2]  Ming-Hsuan Yang,et al.  Kernel Eigenfaces vs. Kernel Fisherfaces: Face recognition using kernel methods , 2002, Proceedings of Fifth IEEE International Conference on Automatic Face Gesture Recognition.

[3]  George Nagy,et al.  Decoding Substitution Ciphers by Means of Word Matching with Application to OCR , 1987, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[4]  Nello Cristianini,et al.  Kernel Methods for Pattern Analysis , 2003, ICTAI.

[5]  H J Caulfield,et al.  Improved discrimination in optical character recognition. , 1969, Applied optics.

[6]  Stephen P. Boyd,et al.  Semidefinite Programming , 1996, SIAM Rev..

[7]  Bernhard Schölkopf,et al.  Nonlinear Component Analysis as a Kernel Eigenvalue Problem , 1998, Neural Computation.

[8]  S T Roweis,et al.  Nonlinear dimensionality reduction by locally linear embedding. , 2000, Science.

[9]  G. Rong,et al.  Learning a data-dependent kernel function for KPCA-based nonlinear process monitoring , 2009 .

[10]  J. Tenenbaum,et al.  A global geometric framework for nonlinear dimensionality reduction. , 2000, Science.

[11]  Kilian Q. Weinberger,et al.  Unsupervised Learning of Image Manifolds by Semidefinite Programming , 2004, Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2004. CVPR 2004..

[12]  Konstantinos N. Plataniotis,et al.  Face recognition using kernel direct discriminant analysis algorithms , 2003, IEEE Trans. Neural Networks.

[13]  Bernhard Schölkopf,et al.  Extracting Support Data for a Given Task , 1995, KDD.

[14]  H. Zha,et al.  Principal manifolds and nonlinear dimensionality reduction via tangent space alignment , 2004, SIAM J. Sci. Comput..

[15]  Thomas S. Huang,et al.  Image processing , 1971 .

[16]  D. Donoho,et al.  Hessian eigenmaps: Locally linear embedding techniques for high-dimensional data , 2003, Proceedings of the National Academy of Sciences of the United States of America.

[17]  Kilian Q. Weinberger,et al.  Learning a kernel matrix for nonlinear dimensionality reduction , 2004, ICML.