A novel deep learning by combining discriminative model with generative model

Deep learning methods allow a classifier to learn features automatically through multiple layers of training. In a deep learning process, low-level features are abstracted into high-level features. In this paper, we propose a new probabilistic deep learning method that combines a discriminative model, namely, Support Vector Machine (SVM), with a generative model, namely, Gaussian Mixture Model (GMM). Combining the SVM with the GMM, we can represent a new input feature for deeper layer training of uncertain data in current layer construction. Bayesian rule is used to re-represent the output data of the previous layer of the SVM with GMM to serve as the input data for the next deep layer. As a result, deep features are reliably extracted without additional feature extraction efforts, using multiple layers of the SVM with GMM. Experimental results show that the proposed deep structure model allows for an easier classification of the uncertain data through multiple-layer training and it gives more accurate results.

[1]  Philippe Refregier,et al.  PROBABILISTIC APPROACH FOR MULTICLASS CLASSIFICATION WITH NEURAL NETWORKS , 1991 .

[2]  L Nyström,et al.  Statistical Analysis , 2008, Encyclopedia of Social Network Analysis and Mining.

[3]  Geoffrey E. Hinton,et al.  Neighbourhood Components Analysis , 2004, NIPS.

[4]  이상헌,et al.  Deep Belief Networks , 2010, Encyclopedia of Machine Learning.

[5]  T. Moon The expectation-maximization algorithm , 1996, IEEE Signal Process. Mag..

[6]  Yee Whye Teh,et al.  A Fast Learning Algorithm for Deep Belief Nets , 2006, Neural Computation.

[7]  Chih-Jen Lin,et al.  LIBSVM: A library for support vector machines , 2011, TIST.

[8]  John M. Quick,et al.  Statistical Analysis with R , 2010 .

[9]  Gerald Tesauro,et al.  Practical issues in temporal difference learning , 1992, Machine Learning.

[10]  TesauroGerald Practical Issues in Temporal Difference Learning , 1992 .

[11]  Sally A. Goldman,et al.  Computational Learning Theory , 2010, Lecture Notes in Computer Science.

[12]  Vangelis Metsis,et al.  Spam Filtering with Naive Bayes - Which Naive Bayes? , 2006, CEAS.

[13]  G. Casella,et al.  Explaining the Gibbs Sampler , 1992 .

[14]  Minho Lee,et al.  Deep Network with Support Vector Machines , 2013, ICONIP.

[15]  Paul Smolensky,et al.  Information processing in dynamical systems: foundations of harmony theory , 1986 .

[16]  Nicolas Le Roux,et al.  Representational Power of Restricted Boltzmann Machines and Deep Belief Networks , 2008, Neural Computation.

[17]  Fei-Fei Li,et al.  What Does Classifying More Than 10, 000 Image Categories Tell Us? , 2010, ECCV.

[18]  Chih-Jen Lin,et al.  Probability Estimates for Multi-class Classification by Pairwise Coupling , 2003, J. Mach. Learn. Res..

[19]  Yichuan Tang,et al.  Deep Learning using Linear Support Vector Machines , 2013, 1306.0239.

[20]  Douglas A. Reynolds,et al.  Gaussian Mixture Models , 2018, Encyclopedia of Biometrics.

[21]  Minho Lee,et al.  Deep learning of support vector machines with class probability output networks , 2015, Neural Networks.

[22]  Yann LeCun,et al.  Regularization of Neural Networks using DropConnect , 2013, ICML.

[23]  Jason Weston,et al.  Multi-Class Support Vector Machines , 1998 .

[24]  Marco Wiering,et al.  Deep Support Vector Machines for Regression Problems , 2013 .

[25]  Nitish Srivastava,et al.  Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..

[26]  William Nick Street,et al.  Breast Cancer Diagnosis and Prognosis Via Linear Programming , 1995, Oper. Res..

[27]  S. Eddy Hidden Markov models. , 1996, Current opinion in structural biology.