Semi-supervised deep network representation with text information

Network representation learning aims at learning low-dimensional representation for each vertex in a network, which plays an important role in network analysis. Con­ventional shallow models often achieve sub-optimal network representation results for non-linear network characteristics. Most network representation methods merely concentrate on structure but ignore text information related to each node. In the paper, we propose a novel semi-supervised deep model for network representation learning. We adopt a random surfing model to capture the global structure and incorporate text features of vertices based on the PV-DBOW model. The joint similarity between vertices achieved by combining network structure and text information is applied as the unsupervised component. While the first-order proximity in a network is used as the supervised component. By jointly optimizing them, our method can obtain reliable low-dimensional vector representations. The experiments on two real-world networks show that our method outperforms other baselines in the task of multi-class classification of vertices.

[1]  Lise Getoor,et al.  Collective Classification in Network Data , 2008, AI Mag..

[2]  Jeffrey Dean,et al.  Efficient Estimation of Word Representations in Vector Space , 2013, ICLR.

[3]  Deli Zhao,et al.  Network Representation Learning with Rich Text Information , 2015, IJCAI.

[4]  Wenwu Zhu,et al.  Structural Deep Network Embedding , 2016, KDD.

[5]  Babak Nasersharif,et al.  Mapping Mel sub-band energies using Deep belief network for robust speech recognition , 2016, 2016 8th International Symposium on Telecommunications (IST).

[6]  Linyuan Lu,et al.  Link prediction based on local random walk , 2010, 1001.2467.

[7]  Steven Skiena,et al.  DeepWalk: online learning of social representations , 2014, KDD.

[8]  Philip S. Yu,et al.  Collective prediction with latent graphs , 2011, CIKM '11.

[9]  Zhiyuan Liu,et al.  Community-enhanced Network Representation Learning for Network Analysis , 2016, ArXiv.

[10]  Feiping Nie,et al.  Cauchy Graph Embedding , 2011, ICML.

[11]  J. Tenenbaum,et al.  A global geometric framework for nonlinear dimensionality reduction. , 2000, Science.

[12]  Michael I. Jordan,et al.  Latent Dirichlet Allocation , 2001, J. Mach. Learn. Res..

[13]  Chengqi Zhang,et al.  Tri-Party Deep Network Representation , 2016, IJCAI.

[14]  Pascal Vincent,et al.  Stacked Denoising Autoencoders: Learning Useful Representations in a Deep Network with a Local Denoising Criterion , 2010, J. Mach. Learn. Res..

[15]  Matthew D. Zeiler Hierarchical Convolutional Deep Learning in Computer Vision , 2013 .

[16]  Mikhail Belkin,et al.  Laplacian Eigenmaps for Dimensionality Reduction and Data Representation , 2003, Neural Computation.

[17]  Mingzhe Wang,et al.  LINE: Large-scale Information Network Embedding , 2015, WWW.

[18]  Geoffrey E. Hinton,et al.  Reducing the Dimensionality of Data with Neural Networks , 2006, Science.

[19]  Quoc V. Le,et al.  Distributed Representations of Sentences and Documents , 2014, ICML.

[20]  Phil Blunsom,et al.  A Convolutional Neural Network for Modelling Sentences , 2014, ACL.