A Semi-supervised Learning Algorithm on Gaussian Mixture with Automatic Model Selection

In Gaussian mixture modeling, it is crucial to select the number of Gaussians for a sample set, which becomes much more difficult when the overlap in the mixture is larger. Under regularization theory, we aim to solve this problem using a semi-supervised learning algorithm through incorporating pairwise constraints into entropy regularized likelihood (ERL) learning which can make automatic model selection for Gaussian mixture. The simulation experiments further demonstrate that the presented semi-supervised learning algorithm (i.e., the constrained ERL learning algorithm) can automatically detect the number of Gaussians with a good parameter estimation, even when two or more actual Gaussians in the mixture are overlapped at a high degree. Moreover, the constrained ERL learning algorithm leads to some promising results when applied to iris data classification and image database categorization.

[1]  G. Schwarz Estimating the Dimension of a Model , 1978 .

[2]  Zhiwu Lu,et al.  A regularized minimum cross-entropy algorithm on mixtures of experts for time series prediction and curve detection , 2006, Pattern Recognit. Lett..

[3]  Tomer Hertz,et al.  Computing Gaussian Mixture Models with EM Using Equivalence Constraints , 2003, NIPS.

[4]  G. R. Dattatreya Gaussian mixture parameter estimation with known means and unknown class-dependent variances , 2002, Pattern Recognit..

[5]  Jinwen Ma,et al.  Asymptotic convergence properties of the EM algorithm with respect to the overlap in the mixture , 2005, Neurocomputing.

[6]  D. Cox,et al.  Asymptotic Analysis of Penalized Likelihood and Related Estimators , 1990 .

[7]  Yi Li,et al.  A generative/discriminative learning algorithm for image classification , 2005, Tenth IEEE International Conference on Computer Vision (ICCV'05) Volume 1.

[8]  Zhiwu Lu An iterative algorithm for entropy regularized likelihood learning on Gaussian mixture with automatic model selection , 2006, Neurocomputing.

[9]  R. Redner,et al.  Mixture densities, maximum likelihood, and the EM algorithm , 1984 .

[10]  H. Akaike A new look at the statistical model identification , 1974 .

[11]  B. S. Manjunath,et al.  Texture Features for Browsing and Retrieval of Image Data , 1996, IEEE Trans. Pattern Anal. Mach. Intell..

[12]  Zhiwu Lu Entropy Regularized Likelihood Learning on Gaussian Mixture: Two Gradient Implementations for Automatic Model Selection , 2006, Neural Processing Letters.