Discriminative Training of Subspace Gaussian Mixture Model for Pattern Classification

The Gaussian mixture model (GMM) has been widely used in pattern recognition problems for clustering and probability density estimation. For pattern classification, however, the GMM has to consider two issues: model structure in high-dimensional space and discriminative training for optimizing the decision boundary. In this paper, we propose a classification method using subspace GMM density model and discriminative training. During discriminative training under the minimum classification error (MCE) criterion, both the GMM parameters and the subspace parameters are optimized discriminatively. Our experimental results on the MNIST handwritten digit data and UCI datasets demonstrate the superior classification performance of the proposed method.

[1]  Yunde Jia,et al.  Unsupervised Selection and Discriminative Estimation of Orthogonal Gaussian Mixture Models for Handwritten Digit Recognition , 2009, 2009 10th International Conference on Document Analysis and Recognition.

[2]  Rui Zhang,et al.  Offline handwritten numeral recognition using orthogonal Gaussian mixture model , 2001, Proceedings 2001 International Conference on Image Processing (Cat. No.01CH37205).

[3]  Biing-Hwang Juang,et al.  Minimum classification error rate methods for speech recognition , 1997, IEEE Trans. Speech Audio Process..

[4]  D. Rubin,et al.  Maximum likelihood from incomplete data via the EM - algorithm plus discussions on the paper , 1977 .

[5]  Mark J. F. Gales,et al.  Semi-tied covariance matrices for hidden Markov models , 1999, IEEE Trans. Speech Audio Process..

[6]  X. Zhou,et al.  Optimisation of Gaussian mixture model for satellite image classification , 2006 .

[7]  Alex Pentland,et al.  Bayesian face recognition , 2000, Pattern Recognit..

[8]  Anil K. Jain,et al.  Unsupervised Learning of Finite Mixture Models , 2002, IEEE Trans. Pattern Anal. Mach. Intell..

[9]  Cheng-Lin Liu,et al.  Handwritten digit recognition: benchmarking of state-of-the-art techniques , 2003, Pattern Recognit..

[10]  Alex Pentland,et al.  Probabilistic Visual Learning for Object Representation , 1997, IEEE Trans. Pattern Anal. Mach. Intell..

[11]  Yoshua Bengio,et al.  Gradient-based learning applied to document recognition , 1998, Proc. IEEE.

[12]  Cheng-Lin Liu,et al.  A pooled subspace mixture density model for pattern classification in high-dimensional spaces , 2008, 2008 IEEE International Joint Conference on Neural Networks (IEEE World Congress on Computational Intelligence).

[13]  Hermann Ney,et al.  Comparison of discriminative training criteria and optimization methods for speech recognition , 2001, Speech Commun..

[14]  Shigeru Katagiri,et al.  Subspace Method for Minimum Error Pattern Recognition , 1997 .

[15]  Hermann Ney,et al.  Discriminative Training of Gaussian Mixtures for Image Object Recognition , 1999, DAGM-Symposium.

[16]  Biing-Hwang Juang,et al.  A new algorithm for fast discriminative training , 2002, 2002 IEEE International Conference on Acoustics, Speech, and Signal Processing.

[17]  Lalit R. Bahl,et al.  Maximum mutual information estimation of hidden Markov model parameters for speech recognition , 1986, ICASSP '86. IEEE International Conference on Acoustics, Speech, and Signal Processing.

[18]  Hiroshi Sako,et al.  Discriminative learning quadratic discriminant function for handwriting recognition , 2004, IEEE Transactions on Neural Networks.