Pointer Graph Networks

Graph neural networks (GNNs) are typically applied to static graphs that are assumed to be known upfront. This static input structure is often informed purely by insight of the machine learning practitioner, and might not be optimal for the actual task the GNN is solving. In absence of reliable domain expertise, one might resort to inferring the latent graph structure, which is often difficult due to the vast search space of possible graphs. Here we introduce Pointer Graph Networks (PGNs) which augment sets or graphs with additional inferred edges for improved model expressivity. PGNs allow each node to dynamically point to another node, followed by message passing over these pointers. The sparsity of this adaptable graph structure makes learning tractable while still being sufficiently expressive to simulate complex algorithms. Critically, the pointing mechanism is directly supervised to model long-term sequences of operations on classical data structures, incorporating useful structural inductive biases from theoretical computer science. Qualitatively, we demonstrate that PGNs can learn parallelisable variants of pointer-based data structures, namely disjoint set unions and link/cut trees. PGNs generalise out-of-distribution to 5x larger test inputs on dynamic graph connectivity tasks, outperforming unrestricted GNNs and Deep Sets.

[1]  Razvan Pascanu,et al.  A simple neural network module for relational reasoning , 2017, NIPS.

[2]  Lukasz Kaiser,et al.  Reformer: The Efficient Transformer , 2020, ICLR.

[3]  Jian Tang,et al.  Graph neural induction of value iteration , 2020, ArXiv.

[4]  Wojciech Zaremba,et al.  Learning to Execute , 2014, ArXiv.

[5]  Shuiwang Ji,et al.  Non-Local Graph Neural Networks , 2020, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[6]  Jure Leskovec,et al.  Improving Graph Attention Networks with Large Margin-based Constraints , 2019, ArXiv.

[7]  E. A. Dinic Algorithm for solution of a problem of maximal flow in a network with power estimation , 1970 .

[8]  Nassir Navab,et al.  Differentiable Graph Module (DGM) for Graph Convolutional Networks , 2020, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[9]  Massimiliano Pontil,et al.  Learning Discrete Structures for Graph Neural Networks , 2019, ICML.

[10]  Shimon Even,et al.  An On-Line Edge-Deletion Problem , 1981, JACM.

[11]  Lukasz Kaiser,et al.  Attention is All you Need , 2017, NIPS.

[12]  Ken-ichi Kawarabayashi,et al.  What Can Neural Networks Reason About? , 2019, ICLR.

[13]  Xin-She Yang,et al.  Introduction to Algorithms , 2021, Nature-Inspired Optimization Algorithms.

[14]  Yoshua Bengio,et al.  Benchmarking Graph Neural Networks , 2023, J. Mach. Learn. Res..

[15]  Demis Hassabis,et al.  Neural Episodic Control , 2017, ICML.

[16]  L. Akoglu,et al.  PairNorm: Tackling Oversmoothing in GNNs , 2019, ICLR.

[17]  Stefanie Jegelka,et al.  Generalization and Representational Limits of Graph Neural Networks , 2020, ICML.

[18]  Michael J. Fischer,et al.  An improved equivalence algorithm , 1964, CACM.

[19]  Robert E. Tarjan,et al.  Efficiency of a Good But Not Linear Set Union Algorithm , 1972, JACM.

[20]  Pietro Li'o,et al.  Neural Bipartite Matching , 2020, ArXiv.

[21]  Jessica B. Hamrick,et al.  Relational inductive bias for physical construction in humans and machines , 2018, CogSci.

[22]  Joan Bruna,et al.  Can graph neural networks count substructures? , 2020, NeurIPS.

[23]  Robert E. Tarjan,et al.  Dynamic trees in practice , 2007, JEAL.

[24]  Lukasz Kaiser,et al.  Neural GPUs Learn Algorithms , 2015, ICLR.

[25]  Chris Dyer,et al.  Neural Arithmetic Logic Units , 2018, NeurIPS.

[26]  Jure Leskovec,et al.  How Powerful are Graph Neural Networks? , 2018, ICLR.

[27]  Danai Koutra,et al.  NEURAL EXECUTION ENGINES , 2019 .

[28]  Jan van Leeuwen,et al.  Worst-case Analysis of Set Union Algorithms , 1984, JACM.

[29]  Roger Wattenhofer,et al.  Normalized Attention Without Probability Cage , 2020, ArXiv.

[30]  Yaron Lipman,et al.  Set2Graph: Learning Graphs From Sets , 2020, NeurIPS.

[31]  Robert E. Tarjan,et al.  Disjoint Set Union with Randomized Linking , 2014, SODA.

[32]  Bao-Liang Lu,et al.  Towards Scale-Invariant Graph-related Problem Solving by Iterative Homogeneous GNNs , 2020, NeurIPS.

[33]  Xu Sun,et al.  Measuring and Relieving the Over-smoothing Problem for Graph Neural Networks from the Topological View , 2019, AAAI.

[34]  Ryan L. Murphy,et al.  Janossy Pooling: Learning Deep Permutation-Invariant Functions for Variable-Size Inputs , 2018, ICLR.

[35]  Bin Luo,et al.  Semi-Supervised Learning With Graph Learning-Convolutional Networks , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[36]  Navdeep Jaitly,et al.  Pointer Networks , 2015, NIPS.

[37]  Alexander J. Smola,et al.  Deep Sets , 2017, 1703.06114.

[38]  R. Hadsell,et al.  Neural Execution of Graph Algorithms , 2019, ICLR.

[39]  J. Kruskal On the shortest spanning subtree of a graph and the traveling salesman problem , 1956 .

[40]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[41]  Razvan Pascanu,et al.  Interaction Networks for Learning about Objects, Relations and Physics , 2016, NIPS.

[42]  Kevin Swersky,et al.  Neural Execution Engines: Learning to Execute Subroutines , 2020, NeurIPS.

[43]  Yuxiao Dong,et al.  DeepInf : Modeling Influence Locality in Large Social Networks , 2018 .

[44]  Robert E. Tarjan,et al.  A data structure for dynamic trees , 1981, STOC '81.

[45]  Richard P. Stanley Acyclic orientations of graphs , 1973, Discret. Math..

[46]  Jure Leskovec,et al.  Learning to Simulate Complex Physics with Graph Networks , 2020, ICML.

[47]  Doina Precup,et al.  Break the Ceiling: Stronger Multi-scale Deep Graph Convolutional Networks , 2019, NeurIPS.

[48]  Samy Bengio,et al.  Order Matters: Sequence to sequence for sets , 2015, ICLR.

[49]  Michael E. Saks,et al.  The cell probe complexity of dynamic data structures , 1989, STOC '89.

[50]  Yue Wang,et al.  Dynamic Graph CNN for Learning on Point Clouds , 2018, ACM Trans. Graph..

[51]  Samuel S. Schoenholz,et al.  Neural Message Passing for Quantum Chemistry , 2017, ICML.

[52]  Max Welling,et al.  Attention, Learn to Solve Routing Problems! , 2018, ICLR.

[53]  Stefano Ermon,et al.  Graphite: Iterative Generative Modeling of Graphs , 2018, ICML.

[54]  Razvan Pascanu,et al.  Learning Deep Generative Models of Graphs , 2018, ICLR 2018.

[55]  R. Zemel,et al.  Neural Relational Inference for Interacting Systems , 2018, ICML.