Using weak supervision in learning Gaussian mixture models

The expectation maximization algorithm is a popular approach to learning Gaussian mixture models from unlabeled data. In addition to the unlabeled data, in many applications, additional sources of information such as a-priori knowledge of mixing proportions are also available. We present a weakly supervised approach, in the form of a penalized expectation maximization algorithm that uses a-priori knowledge to guide the model training process. The algorithm penalizes those models whose predicted mixing proportions have high divergence from the a-priori mixing proportions. We also present an extension to incorporate both labeled and unlabeled data in a semi-supervised setting. Systematic evaluations on several publicly available datasets show that the proposed algorithms outperforms the expectation maximization algorithm. The performance gains are particularly significant when the amount of unlabeled data is limited and in the presence of noise.

[1]  Narendra Ahuja,et al.  Gaussian mixture model for human skin color and its applications in image and video databases , 1998, Electronic Imaging.

[2]  Jun'ichi Tsujii,et al.  Training a Naive Bayes Classifier via the EM Algorithm with a Class Distribution Constraint , 2003, CoNLL.

[3]  Tomer Hertz,et al.  Computing Gaussian Mixture Models with EM Using Equivalence Constraints , 2003, NIPS.

[4]  Douglas A. Reynolds,et al.  Speaker Verification Using Adapted Gaussian Mixture Models , 2000, Digit. Signal Process..

[5]  Qi Zhao,et al.  Semisupervised learning of mixture models with class constraints , 2005, Proceedings. (ICASSP '05). IEEE International Conference on Acoustics, Speech, and Signal Processing, 2005..

[6]  Claire Cardie,et al.  Constrained K-means Clustering with Background Knowledge , 2001, ICML.

[7]  Christopher M. Bishop,et al.  Pattern Recognition and Machine Learning (Information Science and Statistics) , 2006 .

[8]  Tao Jiang,et al.  The Regularized EM Algorithm , 2005, AAAI.

[9]  Sebastian Thrun,et al.  Text Classification from Labeled and Unlabeled Documents using EM , 2000, Machine Learning.

[10]  Geoffrey J. McLachlan,et al.  Mixture models : inference and applications to clustering , 1989 .

[11]  Michael I. Jordan,et al.  Mixture models for learning from incomplete data , 1997, COLT 1997.

[12]  N. Bing,et al.  Finite mixture model analysis of microarray expression data on samples of uncertain biological type with application to reproductive efficiency. , 2005, Veterinary immunology and immunopathology.

[13]  Patrice Loisel,et al.  Testing in normal mixture models when the proportions are known , 1992 .

[14]  Gideon S. Mann,et al.  Simple, robust, scalable semi-supervised learning via expectation regularization , 2007, ICML '07.

[15]  Alexander J. Smola,et al.  Estimating labels from label proportions , 2008, ICML '08.

[16]  Nicu Sebe,et al.  Semisupervised learning of classifiers: theory, algorithms, and their application to human-computer interaction , 2004, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[17]  David A. Landgrebe,et al.  Classification of multispectral data by joint supervised-unsupervised learning , 1993 .

[18]  Radford M. Neal Pattern Recognition and Machine Learning , 2007, Technometrics.