Transformer-Style Relational Reasoning with Dynamic Memory Updating for Temporal Network Modeling

Network modeling aims to learn the latent representations of nodes such that the representations preserve both network structures and node attribute information. This problem is fundamental due to its prevalence in numerous domains. However, existing approaches either target the static networks or struggle to capture the complicated temporal dependency, while most real-world networks evolve over time and the success of network modeling hinges on the understanding of how entities are temporally connected. In this paper, we present TRRN, a transformer-style relational reasoning network with dynamic memory updating, to deal with the above challenges. TRRN employs multi-head self-attention to reason over a set of memories, which provides a multitude of shortcut paths for information to flow from past observations to the current latent representations. By utilizing the policy networks augmented with differentiable binary routers, TRRN estimates the possibility of each memory being activated and dynamically updates the memories at the time steps when they are most relevant. We evaluate TRRN with the tasks of node classification and link prediction on four real temporal network datasets. Experimental results demonstrate the consistent performance gains for TRRN over the leading competitors.

[1]  Jian Pei,et al.  Community Preserving Network Embedding , 2017, AAAI.

[2]  Xiang Zhang,et al.  Spatio-Temporal Attentive RNN for Node Classification in Temporal Attributed Graphs , 2019, IJCAI.

[3]  Joelle Pineau,et al.  Focused Hierarchical RNNs for Conditional Sequence Processing , 2018, ICML.

[4]  Jure Leskovec,et al.  Inductive Representation Learning on Large Graphs , 2017, NIPS.

[5]  Pietro Liò,et al.  Graph Attention Networks , 2017, ICLR.

[6]  Palash Goyal,et al.  dyngraph2vec: Capturing Network Dynamics using Dynamic Graph Representation Learning , 2018, Knowl. Based Syst..

[7]  Bo Zong,et al.  Adaptive Neural Network for Node Classification in Dynamic Networks , 2019, 2019 IEEE International Conference on Data Mining (ICDM).

[8]  Liang Gou,et al.  DySAT: Deep Neural Representation Learning on Dynamic Graphs via Self-Attention Networks , 2020, WSDM.

[9]  Lorenzo Torresani,et al.  STAR-Caps: Capsule Networks with Straight-Through Attentive Routing , 2019, NeurIPS.

[10]  Max Welling,et al.  Semi-Supervised Classification with Graph Convolutional Networks , 2016, ICLR.

[11]  Yoshua Bengio,et al.  The Consciousness Prior , 2017, ArXiv.

[12]  Guojie Song,et al.  Dynamic Network Embedding : An Extended Approach for Skip-gram based Network Embedding , 2018, IJCAI.

[13]  Jeffrey Dean,et al.  Efficient Estimation of Word Representations in Vector Space , 2013, ICLR.

[14]  Nicholas Jing Yuan,et al.  Integrating Graph Contextualized Knowledge into Pre-trained Language Models , 2019, FINDINGS.

[15]  Razvan Pascanu,et al.  Relational inductive biases, deep learning, and graph networks , 2018, ArXiv.

[16]  Lukasz Kaiser,et al.  Attention is All you Need , 2017, NIPS.

[17]  Razvan Pascanu,et al.  Relational recurrent neural networks , 2018, NeurIPS.

[18]  Yee Whye Teh,et al.  The Concrete Distribution: A Continuous Relaxation of Discrete Random Variables , 2016, ICLR.

[19]  Ben Poole,et al.  Categorical Reparameterization with Gumbel-Softmax , 2016, ICLR.

[20]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[21]  Laura C. Buchanan,et al.  Tracking ongoing cognition in individuals using brief, whole-brain functional connectivity patterns , 2015, Proceedings of the National Academy of Sciences.

[22]  Jian Pei,et al.  High-Order Proximity Preserved Embedding for Dynamic Networks , 2018, IEEE Transactions on Knowledge and Data Engineering.

[23]  Zhiyuan Liu,et al.  Graph Neural Networks: A Review of Methods and Applications , 2018, AI Open.

[24]  Wei Lu,et al.  Deep Neural Networks for Learning Graph Representations , 2016, AAAI.

[25]  Yoshua Bengio,et al.  Estimating or Propagating Gradients Through Stochastic Neurons for Conditional Computation , 2013, ArXiv.

[26]  Rogério Schmidt Feris,et al.  SpotTune: Transfer Learning Through Adaptive Fine-Tuning , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[27]  Yan Liu,et al.  DynGEM: Deep Embedding Method for Dynamic Graphs , 2018, ArXiv.

[28]  Huan Liu,et al.  Attributed Network Embedding for Learning in a Dynamic Environment , 2017, CIKM.

[29]  Wenwu Zhu,et al.  Deep Learning on Graphs: A Survey , 2018, IEEE Transactions on Knowledge and Data Engineering.

[30]  Jiliang Tang,et al.  Streaming Graph Neural Networks , 2018, SIGIR.

[31]  Bernhard Schölkopf,et al.  Recurrent Independent Mechanisms , 2021, ICLR.

[32]  Junjie Wu,et al.  Embedding Temporal Network via Neighborhood Formation , 2018, KDD.

[33]  Philip S. Yu,et al.  A Comprehensive Survey on Graph Neural Networks , 2019, IEEE Transactions on Neural Networks and Learning Systems.

[34]  Wenwu Zhu,et al.  Structural Deep Network Embedding , 2016, KDD.

[35]  Steven Skiena,et al.  DeepWalk: online learning of social representations , 2014, KDD.

[36]  Jure Leskovec,et al.  node2vec: Scalable Feature Learning for Networks , 2016, KDD.

[37]  Jürgen Schmidhuber,et al.  Long Short-Term Memory , 1997, Neural Computation.

[38]  Sergio Gomez Colmenarejo,et al.  Hybrid computing using a neural network with dynamic external memory , 2016, Nature.

[39]  Yoshua Bengio,et al.  Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation , 2014, EMNLP.

[40]  Daan Wierstra,et al.  Meta-Learning with Memory-Augmented Neural Networks , 2016, ICML.

[41]  Jason Weston,et al.  End-To-End Memory Networks , 2015, NIPS.