Pointer Graph Networks

Graph neural networks (GNNs) are typically applied to static graphs that are assumed to be known upfront. This static input structure is often informed purely by insight of the machine learning practitioner, and might not be optimal for the actual task the GNN is solving. In absence of reliable domain expertise, one might resort to inferring the latent graph structure, which is often difficult due to the vast search space of possible graphs. Here we introduce Pointer Graph Networks (PGNs) which augment sets or graphs with additional inferred edges for improved model expressivity. PGNs allow each node to dynamically point to another node, followed by message passing over these pointers. The sparsity of this adaptable graph structure makes learning tractable while still being sufficiently expressive to simulate complex algorithms. Critically, the pointing mechanism is directly supervised to model long-term sequences of operations on classical data structures, incorporating useful structural inductive biases from theoretical computer science. Qualitatively, we demonstrate that PGNs can learn parallelisable variants of pointer-based data structures, namely disjoint set unions and link/cut trees. PGNs generalise out-of-distribution to 5x larger test inputs on dynamic graph connectivity tasks, outperforming unrestricted GNNs and Deep Sets.

[1]  Joan Bruna,et al.  Can graph neural networks count substructures? , 2020, NeurIPS.

[2]  R. Zemel,et al.  Neural Relational Inference for Interacting Systems , 2018, ICML.

[3]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[4]  Alexander J. Smola,et al.  Deep Sets , 2017, 1703.06114.

[5]  Lukasz Kaiser,et al.  Reformer: The Efficient Transformer , 2020, ICLR.

[6]  Michael E. Saks,et al.  The cell probe complexity of dynamic data structures , 1989, STOC '89.

[7]  Jian Tang,et al.  Graph neural induction of value iteration , 2020, ArXiv.

[8]  Massimiliano Pontil,et al.  Learning Discrete Structures for Graph Neural Networks , 2019, ICML.

[9]  Yuxiao Dong,et al.  DeepInf : Modeling Influence Locality in Large Social Networks , 2018 .

[10]  Wojciech Zaremba,et al.  Learning to Execute , 2014, ArXiv.

[11]  Yaron Lipman,et al.  Set2Graph: Learning Graphs From Sets , 2020, NeurIPS.

[12]  Bin Luo,et al.  Semi-Supervised Learning With Graph Learning-Convolutional Networks , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[13]  Samuel S. Schoenholz,et al.  Neural Message Passing for Quantum Chemistry , 2017, ICML.

[14]  Max Welling,et al.  Attention, Learn to Solve Routing Problems! , 2018, ICLR.

[15]  Shimon Even,et al.  An On-Line Edge-Deletion Problem , 1981, JACM.

[16]  Razvan Pascanu,et al.  Learning Deep Generative Models of Graphs , 2018, ICLR 2018.

[17]  Pietro Li'o,et al.  Neural Bipartite Matching , 2020, ArXiv.

[18]  Raia Hadsell,et al.  Neural Execution of Graph Algorithms , 2020, ICLR.

[19]  Bao-Liang Lu,et al.  Towards Scale-Invariant Graph-related Problem Solving by Iterative Homogeneous GNNs , 2020, NeurIPS.

[20]  Robert E. Tarjan,et al.  A data structure for dynamic trees , 1981, STOC '81.

[21]  Danai Koutra,et al.  Neural Execution Engines: Learning to Execute Subroutines , 2020, NeurIPS.

[22]  Robert E. Tarjan,et al.  Efficiency of a Good But Not Linear Set Union Algorithm , 1972, JACM.

[23]  Jure Leskovec,et al.  Improving Graph Attention Networks with Large Margin-based Constraints , 2019, ArXiv.

[24]  Stefanie Jegelka,et al.  Generalization and Representational Limits of Graph Neural Networks , 2020, ICML.

[25]  Doina Precup,et al.  Break the Ceiling: Stronger Multi-scale Deep Graph Convolutional Networks , 2019, NeurIPS.

[26]  Jan van Leeuwen,et al.  Worst-case Analysis of Set Union Algorithms , 1984, JACM.

[27]  Jessica B. Hamrick,et al.  Relational inductive bias for physical construction in humans and machines , 2018, CogSci.

[28]  J. Kruskal On the shortest spanning subtree of a graph and the traveling salesman problem , 1956 .

[29]  Ken-ichi Kawarabayashi,et al.  What Can Neural Networks Reason About? , 2019, ICLR.

[30]  Razvan Pascanu,et al.  A simple neural network module for relational reasoning , 2017, NIPS.

[31]  Shuiwang Ji,et al.  Non-Local Graph Neural Networks , 2020, ArXiv.

[32]  Razvan Pascanu,et al.  Interaction Networks for Learning about Objects, Relations and Physics , 2016, NIPS.

[33]  Navdeep Jaitly,et al.  Pointer Networks , 2015, NIPS.

[34]  Richard P. Stanley Acyclic orientations of graphs , 1973, Discret. Math..

[35]  Jure Leskovec,et al.  Learning to Simulate Complex Physics with Graph Networks , 2020, ICML.

[36]  Lukasz Kaiser,et al.  Neural GPUs Learn Algorithms , 2015, ICLR.

[37]  Danai Koutra,et al.  NEURAL EXECUTION ENGINES , 2019 .

[38]  Robert E. Tarjan,et al.  Finding Biconnected Components and Computing Tree Functions in Logarithmic Parallel Time (Extended Summary) , 1984, FOCS.

[39]  Lukasz Kaiser,et al.  Attention is All you Need , 2017, NIPS.

[40]  Robert E. Tarjan,et al.  Disjoint Set Union with Randomized Linking , 2014, SODA.

[41]  Yue Wang,et al.  Dynamic Graph CNN for Learning on Point Clouds , 2018, ACM Trans. Graph..

[42]  Xin-She Yang,et al.  Introduction to Algorithms , 2021, Nature-Inspired Optimization Algorithms.

[43]  Roger Wattenhofer,et al.  Normalized Attention Without Probability Cage , 2020, ArXiv.

[44]  Yoshua Bengio,et al.  Benchmarking Graph Neural Networks , 2023, J. Mach. Learn. Res..

[45]  Robert E. Tarjan,et al.  Dynamic trees in practice , 2007, JEAL.

[46]  Samy Bengio,et al.  Order Matters: Sequence to sequence for sets , 2015, ICLR.

[47]  E. A. Dinic Algorithm for solution of a problem of maximal flow in a network with power estimation , 1970 .

[48]  Nassir Navab,et al.  Differentiable Graph Module (DGM) for Graph Convolutional Networks , 2020, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[49]  Jure Leskovec,et al.  How Powerful are Graph Neural Networks? , 2018, ICLR.

[50]  Chris Dyer,et al.  Neural Arithmetic Logic Units , 2018, NeurIPS.

[51]  Ryan L. Murphy,et al.  Janossy Pooling: Learning Deep Permutation-Invariant Functions for Variable-Size Inputs , 2018, ICLR.

[52]  Demis Hassabis,et al.  Neural Episodic Control , 2017, ICML.

[53]  Jie Zhou,et al.  Measuring and Relieving the Over-smoothing Problem for Graph Neural Networks from the Topological View , 2020, AAAI.

[54]  Michael J. Fischer,et al.  An improved equivalence algorithm , 1964, CACM.

[55]  Stefano Ermon,et al.  Graphite: Iterative Generative Modeling of Graphs , 2018, ICML.