Towards Open Temporal Graph Neural Networks

Graph neural networks (GNNs) for temporal graphs have recently attracted increasing attentions, where a common assumption is that the class set for nodes is closed. However, in real-world scenarios, it often faces the open set problem with the dynamically increased class set as the time passes by. This will bring two big challenges to the existing dynamic GNN methods: (i) How to dynamically propagate appropriate information in an open temporal graph, where new class nodes are often linked to old class nodes. This case will lead to a sharp contradiction. This is because typical GNNs are prone to make the embeddings of connected nodes become similar, while we expect the embeddings of these two interactive nodes to be distinguishable since they belong to different classes. (ii) How to avoid catastrophic knowledge forgetting over old classes when learning new classes occurred in temporal graphs. In this paper, we propose a general and principled learning approach for open temporal graphs, called OTGNet, with the goal of addressing the above two challenges. We assume the knowledge of a node can be disentangled into class-relevant and class-agnostic one, and thus explore a new message passing mechanism by extending the information bottleneck principle to only propagate class-agnostic knowledge between nodes of different classes, avoiding aggregating conflictive information. Moreover, we devise a strategy to select both important and diverse triad sub-graph structures for effective class-incremental learning. Extensive experiments on three real-world datasets of different domains demonstrate the superiority of our method, compared to the baselines.

[1]  H. Zha,et al.  Robust Knowledge Adaptation for Dynamic Graph Neural Networks , 2022, ArXiv.

[2]  Guoren Wang,et al.  FreeKD: Free-direction Knowledge Distillation for Graph Neural Networks , 2022, KDD.

[3]  Yuan Fang,et al.  TREND: TempoRal Event and Node Dynamics for Graph Representation Learning , 2022, WWW.

[4]  Dongmei Zhang,et al.  GBK-GNN: Gated Bi-Kernel Graph Neural Networks for Modeling Both Homophily and Heterophily , 2021, WWW.

[5]  Fan Zhou,et al.  Overcoming Catastrophic Forgetting in Graph Neural Networks with Experience Replay , 2021, AAAI.

[6]  Qinghua Hu,et al.  Multi-View Information-Bottleneck Representation Learning , 2021, AAAI.

[7]  Jihwan Bang,et al.  Rainbow Memory: Continual Learning with a Memory of Diverse Samples , 2021, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[8]  Jure Leskovec,et al.  Inductive Representation Learning in Temporal Networks via Causal Anonymous Walks , 2021, ICLR.

[9]  Yiding Yang,et al.  Overcoming Catastrophic Forgetting in Graph Neural Networks , 2020, AAAI.

[10]  Yupu Yang,et al.  APAN: Asynchronous Propagation Attention Network for Real-time Temporal Graph Embedding , 2020, SIGMOD Conference.

[11]  Yi Wu,et al.  Streaming Graph Neural Networks via Continual Learning , 2020, CIKM.

[12]  Raymond Chi-Wing Wong,et al.  When Do GNNs Work: Understanding and Improving Neighborhood Aggregation , 2020, IJCAI.

[13]  Zhe Gan,et al.  CLUB: A Contrastive Log-ratio Upper Bound of Mutual Information , 2020, ICML.

[14]  L. Akoglu,et al.  Beyond Homophily in Graph Neural Networks: Current Limitations and Effective Designs , 2020, NeurIPS.

[15]  Davide Eynard,et al.  Temporal Graph Networks for Deep Learning on Dynamic Graphs , 2020, ArXiv.

[16]  Xiaopeng Hong,et al.  Few-Shot Class-Incremental Learning , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[17]  Xiaoyang Wang,et al.  Traffic Flow Prediction via Spatial Temporal Graph Neural Network , 2020, WWW.

[18]  Da Xu,et al.  Inductive Representation Learning on Temporal Graphs , 2020, ICLR.

[19]  Mohit Singh,et al.  On the Unreasonable Effectiveness of the Greedy Algorithm: Greedy Adapts to Sharpness , 2020, ICML 2020.

[20]  Liang Gou,et al.  DySAT: Deep Neural Representation Learning on Dynamic Graphs via Self-Attention Networks , 2020, WSDM.

[21]  Jure Leskovec,et al.  Predicting Dynamic Embedding Trajectory in Temporal Interaction Networks , 2019, KDD.

[22]  Xiaowei Wang,et al.  Sequential Scenario-Specific Meta Learner for Online Recommendation , 2019, KDD.

[23]  Yandong Guo,et al.  Large Scale Incremental Learning , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[24]  Hongyuan Zha,et al.  DyRep: Learning Representations over Dynamic Graphs , 2019, ICLR.

[25]  Yan Liu,et al.  DynGEM: Deep Embedding Method for Dynamic Graphs , 2018, ArXiv.

[26]  Yueting Zhuang,et al.  Dynamic Network Embedding by Modeling Triadic Closure Process , 2018, AAAI.

[27]  Ryan A. Rossi,et al.  Continuous-Time Dynamic Network Embeddings , 2018, WWW.

[28]  Philip H. S. Torr,et al.  Riemannian Walk for Incremental Learning: Understanding Forgetting and Intransigence , 2018, ECCV.

[29]  Aaron C. Courville,et al.  Mutual Information Neural Estimation , 2018, ICML.

[30]  Jure Leskovec,et al.  Inductive Representation Learning on Large Graphs , 2017, NIPS.

[31]  Percy Liang,et al.  Understanding Black-box Predictions via Influence Functions , 2017, ICML.

[32]  Andrei A. Rusu,et al.  Overcoming catastrophic forgetting in neural networks , 2016, Proceedings of the National Academy of Sciences.

[33]  Christoph H. Lampert,et al.  iCaRL: Incremental Classifier and Representation Learning , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[34]  Derek Hoiem,et al.  Learning without Forgetting , 2016, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[35]  M. Kian Operator Jensen inequality for superquadratic functions , 2014 .

[36]  Ben Y. Zhao,et al.  Link and Triadic Closure Delay: Temporal Metrics for Social Network Dynamics , 2014, ICWSM.

[37]  Xiaoming Fu,et al.  Mining triadic closure patterns in social networks , 2014, WWW.

[38]  Jon M. Kleinberg,et al.  Challenges in mining social network data: processes, privacy, and paradoxes , 2007, KDD '07.

[39]  Naftali Tishby,et al.  The information bottleneck method , 2000, ArXiv.

[40]  S. Weisberg,et al.  Characterizations of an Empirical Influence Function for Detecting Influential Cases in Regression , 1980 .

[41]  Srinivasa Varadhan,et al.  Asymptotics for the wiener sausage , 1975 .

[42]  Andreas Krause,et al.  Submodular Function Maximization , 2014, Tractability.

[43]  Geoffrey E. Hinton,et al.  Visualizing Data using t-SNE , 2008 .