Multimodal correlation deep belief networks for multi-view classification

The Restricted Boltzmann machine (RBM) has been proven to be a powerful tool in many specific applications, such as representational learning, document modeling, and many other learning tasks. However, the extensions of the RBM are rarely used in the field of multi-view learning. In this paper, we present a new RBM model based on canonical correlation analysis, named as the correlation RBM, for multi-view learning. The correlation RBM computes multiple representations by regularizing the marginal likelihood function with the consistency among representations from different views. In addition, the multimodal deep model can obtain a unified representation that fuses multiple representations together. Therefore, we stack the correlation RBM to create the correlation deep belief network (DBN), and then propose the multimodal correlation DBN for learning multi-view data representations. Contrasting with existing multi-view classification methods, such as multi-view Gaussian process with posterior consistency (MvGP) and consensus and complementarity based maximum entropy discrimination (MED-2C), the correlation RBM and the multimodal correlation DBN have achieved satisfactory results on two-class and multi-class classification datasets. Experimental results show that correlation RBM and the multimodal correlation DBN are effective learning algorithms.

[1]  P. S. Sastry,et al.  An Overview of Restricted Boltzmann Machines , 2019, Journal of the Indian Institute of Science.

[2]  Shifei Ding,et al.  An overview on Restricted Boltzmann Machines , 2018, Neurocomputing.

[3]  Geoffrey E. Hinton,et al.  Reducing the Dimensionality of Data with Neural Networks , 2006, Science.

[4]  Shiliang Sun,et al.  Multi-view learning overview: Recent progress and new challenges , 2017, Inf. Fusion.

[5]  H. Altay Güvenir,et al.  Learning differential diagnosis of erythemato-squamous diseases using voting feature intervals , 1998, Artif. Intell. Medicine.

[6]  Dale Schuurmans,et al.  Stochastic Neural Networks with Monotonic Activation Functions , 2016, AISTATS.

[7]  Roohallah Alizadehsani,et al.  Computer aided decision making for heart disease detection using hybrid neural network-Genetic algorithm , 2017, Comput. Methods Programs Biomed..

[8]  Hamido Fujita,et al.  A study of graph-based system for multi-view clustering , 2019, Knowl. Based Syst..

[9]  Yu Xue,et al.  Weight Uncertainty in Boltzmann Machine , 2016, Cognitive Computation.

[10]  Shiliang Sun,et al.  Multi-view Regularized Gaussian Processes , 2017, PAKDD.

[11]  Geoffrey E. Hinton,et al.  Deep Boltzmann Machines , 2009, AISTATS.

[12]  Yoshua Bengio,et al.  The Spike-and-Slab RBM and Extensions to Discrete and Sparse Data Distributions , 2014, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[13]  Yu Xue,et al.  Weighted linear loss multiple birth support vector machine based on information granulation for multi-class classification , 2017, Pattern Recognit..

[14]  Sangram Ganguly,et al.  Learning Sparse Feature Representations Using Probabilistic Quadtrees and Deep Belief Nets , 2015, Neural Processing Letters.

[15]  Svetha Venkatesh,et al.  Graph-induced restricted Boltzmann machines for document modeling , 2016, Inf. Sci..

[16]  Ryutaro Tateishi,et al.  Using geographically weighted variables for image classification , 2012 .

[17]  Silvio Savarese,et al.  Structured Recurrent Temporal Restricted Boltzmann Machines , 2014, ICML.

[18]  Zhang Yi,et al.  A multitask multiview clustering algorithm in heterogeneous situations based on LLE and LE , 2019, Knowl. Based Syst..

[19]  Mohamed R. Amer,et al.  Deep Multimodal Fusion: A Hybrid Approach , 2017, International Journal of Computer Vision.

[20]  Shifei Ding,et al.  Research on Point-wise Gated Deep Networks , 2017, Appl. Soft Comput..

[21]  Jeff A. Bilmes,et al.  Deep Canonical Correlation Analysis , 2013, ICML.

[22]  Nitish Srivastava,et al.  Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..

[23]  Chun-Liang Li,et al.  Annealing Gaussian into ReLU: a New Sampling Strategy for Leaky-ReLU RBM , 2016, ArXiv.

[24]  Geoffrey E. Hinton Training Products of Experts by Minimizing Contrastive Divergence , 2002, Neural Computation.

[25]  William Nick Street,et al.  Breast Cancer Diagnosis and Prognosis Via Linear Programming , 1995, Oper. Res..

[26]  Shiliang Sun,et al.  Consensus and complementarity based maximum entropy discrimination for multi-view classification , 2016, Inf. Sci..