Non-Linear Smoothed Transductive Network Embedding with Text Information

Network embedding is a classical task which aims to map the nodes of a network to lowdimensional vectors. Most of the previous network embedding methods are trained in an unsupervised scheme. Then the learned node embeddings can be used as inputs of many machine learning tasks such as node classification, attribute inference. However, the discriminant power of the node embeddings maybe improved by considering the node label information and the node attribute information. Inspired by traditional semi-supervised learning techniques, we explore to train the node embeddings and the node classifiers simultaneously with the text attributes information in a flexible framework. We present Non-Linear Smoothed Transductive Network Embedding (NLSTNE), a transductive network embedding method, whose embeddings are enhanced by modeling the non-linear pairwise similarity between the nodes and the non-linear relationships between the nodes and the text attributes. We use the node classification task to evaluate the quality of the node embeddings learned by different models on four real-world network datasets. The experimental results demonstrate that our model outperforms several state-of-the-art network embedding methods.

[1]  R. A. Leibler,et al.  On Information and Sufficiency , 1951 .

[2]  Jingzhou Liu,et al.  Visualization Large-scale and High-dimensional Data , 2016, arXiv.org.

[3]  Ruslan Salakhutdinov,et al.  Revisiting Semi-Supervised Learning with Graph Embeddings , 2016, ICML.

[4]  Steven Skiena,et al.  Walklets: Multiscale Graph Embeddings for Interpretable Network Classification , 2016, ArXiv.

[5]  Yizhou Sun,et al.  iTopicModel: Information Network-Integrated Topic Modeling , 2009, 2009 Ninth IEEE International Conference on Data Mining.

[6]  Jeffrey Dean,et al.  Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.

[7]  Deli Zhao,et al.  Network Representation Learning with Rich Text Information , 2015, IJCAI.

[8]  Li Liu,et al.  Aligning Users across Social Networks Using Network Embedding , 2016, IJCAI.

[9]  Qiongkai Xu,et al.  GraRep: Learning Graph Representations with Global Structural Information , 2015, CIKM.

[10]  S T Roweis,et al.  Nonlinear dimensionality reduction by locally linear embedding. , 2000, Science.

[11]  Mo Chen,et al.  Directed Graph Embedding , 2007, IJCAI.

[12]  Steven Skiena,et al.  DeepWalk: online learning of social representations , 2014, KDD.

[13]  Nagarajan Natarajan,et al.  Inductive matrix completion for predicting gene–disease associations , 2014, Bioinform..

[14]  Zhiyuan Liu,et al.  Max-Margin DeepWalk: Discriminative Learning of Network Representation , 2016, IJCAI.

[15]  Bo Zhang,et al.  Discriminative Deep Random Walk for Network Classification , 2016, ACL.

[16]  Mingzhe Wang,et al.  LINE: Large-scale Information Network Embedding , 2015, WWW.

[17]  Feiping Nie,et al.  Cauchy Graph Embedding , 2011, ICML.

[18]  A. J. Walker New fast method for generating discrete random numbers with arbitrary frequency distributions , 1974 .

[19]  Ludovic Denoyer,et al.  Learning latent representations of nodes for classifying in heterogeneous social networks , 2014, WSDM.

[20]  Jie Tang,et al.  Transfer Link Prediction across Heterogeneous Social Networks , 2015 .

[21]  Mikhail Belkin,et al.  Laplacian Eigenmaps and Spectral Techniques for Embedding and Clustering , 2001, NIPS.