Provably Powerful Graph Networks

Recently, the Weisfeiler-Lehman (WL) graph isomorphism test was used to measure the expressive power of graph neural networks (GNN). It was shown that the popular message passing GNN cannot distinguish between graphs that are indistinguishable by the 1-WL test (Morris et al. 2018; Xu et al. 2019). Unfortunately, many simple instances of graphs are indistinguishable by the 1-WL test. In search for more expressive graph learning models we build upon the recent k-order invariant and equivariant graph neural networks (Maron et al. 2019a,b) and present two results: First, we show that such k-order networks can distinguish between non-isomorphic graphs as good as the k-WL tests, which are provably stronger than the 1-WL test for k>2. This makes these models strictly stronger than message passing models. Unfortunately, the higher expressiveness of these models comes with a computational cost of processing high order tensors. Second, setting our goal at building a provably stronger, simple and scalable model we show that a reduced 2-order network containing just scaled identity operator, augmented with a single quadratic operation (matrix multiplication) has a provable 3-WL expressive power. Differently put, we suggest a simple model that interleaves applications of standard Multilayer-Perceptron (MLP) applied to the feature dimension and matrix multiplication. We validate this model by presenting state of the art results on popular graph classification and regression tasks. To the best of our knowledge, this is the first practical invariant/equivariant model with guaranteed 3-WL expressiveness, strictly stronger than message passing models.

[1]  Martin Grohe,et al.  Descriptive Complexity, Canonisation, and Definable Graph Structure Theory , 2017, Lecture Notes in Logic.

[2]  S. V. N. Vishwanathan,et al.  Graph kernels , 2007 .

[3]  Kristian Kersting,et al.  Global Weisfeiler-Lehman Graph Kernels , 2017, ArXiv.

[4]  Yixin Chen,et al.  An End-to-End Deep Learning Architecture for Graph Classification , 2018, AAAI.

[5]  Yixin Chen,et al.  Weisfeiler-Lehman Neural Machine for Link Prediction , 2017, KDD.

[6]  Emmanuel Briand,et al.  When is the algebra of multisymmetric polynomials generated by the elementary multisymmetric polynomials , 2004 .

[7]  Sergey Ivanov,et al.  Anonymous Walk Embeddings , 2018, ICML.

[8]  Petra Mutzel,et al.  Towards a practical k-dimensional Weisfeiler-Leman algorithm , 2019, ArXiv.

[9]  Pavlo O. Dral,et al.  Quantum chemistry structures and properties of 134 kilo molecules , 2014, Scientific Data.

[10]  Zhi-Li Zhang,et al.  Hunt For The Unique, Stable, Sparse And Fast Feature Learning On Graphs , 2017, NIPS.

[11]  Christopher Morris,et al.  Weisfeiler and Leman go sparse: Towards scalable higher-order graph embeddings , 2020, NeurIPS.

[12]  Jure Leskovec,et al.  Representation Learning on Graphs: Methods and Applications , 2017, IEEE Data Eng. Bull..

[13]  G. Lewicki,et al.  Approximation by Superpositions of a Sigmoidal Function , 2003 .

[14]  David Rydh,et al.  A minimal Set of Generators for the Ring of multisymmetric Functions , 2007, 0710.0470.

[15]  Jure Leskovec,et al.  Hierarchical Graph Representation Learning with Differentiable Pooling , 2018, NeurIPS.

[16]  Nils M. Kriege,et al.  A survey on graph kernels , 2019, Applied Network Science.

[17]  Kurt Mehlhorn,et al.  Efficient graphlet kernels for large graph comparison , 2009, AISTATS.

[18]  Pinar Yanardag,et al.  Deep Graph Kernels , 2015, KDD.

[19]  Yaron Lipman,et al.  Invariant and Equivariant Graph Networks , 2018, ICLR.

[20]  Vinayak A. Rao,et al.  Relational Pooling for Graph Representations , 2019, ICML.

[21]  Gabriel Peyré,et al.  Universal Invariant and Equivariant Graph Neural Networks , 2019, NeurIPS.

[22]  Nikos Komodakis,et al.  Dynamic Edge-Conditioned Filters in Convolutional Neural Networks on Graphs , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[23]  Jure Leskovec,et al.  How Powerful are Graph Neural Networks? , 2018, ICLR.

[24]  Ah Chung Tsoi,et al.  The Graph Neural Network Model , 2009, IEEE Transactions on Neural Networks.

[25]  Kurt Mehlhorn,et al.  Weisfeiler-Lehman Graph Kernels , 2011, J. Mach. Learn. Res..

[26]  Roman Garnett,et al.  Propagation kernels: efficient graph kernels from propagated information , 2015, Machine Learning.

[27]  Martin Grohe,et al.  Weisfeiler and Leman Go Neural: Higher-order Graph Neural Networks , 2018, AAAI.

[28]  Xavier Bresson,et al.  CayleyNets: Graph Convolutional Neural Networks With Complex Rational Spectral Filters , 2017, IEEE Transactions on Signal Processing.

[29]  Mathias Niepert,et al.  Learning Convolutional Neural Networks for Graphs , 2016, ICML.

[30]  Risi Kondor,et al.  Covariant Compositional Networks For Learning Graphs , 2018, ICLR.

[31]  Joan Bruna,et al.  Deep Convolutional Networks on Graph-Structured Data , 2015, ArXiv.

[32]  Alexander J. Smola,et al.  Deep Sets , 2017, 1703.06114.

[33]  Joan Bruna,et al.  Spectral Networks and Locally Connected Networks on Graphs , 2013, ICLR.

[34]  Yuan Yu,et al.  TensorFlow: A system for large-scale machine learning , 2016, OSDI.

[35]  Neil Immerman,et al.  An optimal lower bound on the number of variables for graph identification , 1992, Comb..

[36]  Samuel S. Schoenholz,et al.  Neural Message Passing for Quantum Chemistry , 2017, ICML.

[37]  B. Douglas The Weisfeiler-Lehman Method and Graph Isomorphism Testing , 2011, 1101.5211.

[38]  Martin Grohe,et al.  PEBBLE GAMES AND LINEAR EQUATIONS , 2012, The Journal of Symbolic Logic.

[39]  Xavier Bresson,et al.  Convolutional Neural Networks on Graphs with Fast Localized Spectral Filtering , 2016, NIPS.

[40]  Yaron Lipman,et al.  On the Universality of Invariant Networks , 2019, ICML.

[41]  Kurt Hornik,et al.  Approximation capabilities of multilayer feedforward networks , 1991, Neural Networks.

[42]  F. Scarselli,et al.  A new model for learning in graph domains , 2005, Proceedings. 2005 IEEE International Joint Conference on Neural Networks, 2005..

[43]  Jonathan Masci,et al.  Geometric Deep Learning on Graphs and Manifolds Using Mixture Model CNNs , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[44]  Donald F. Towsley,et al.  Diffusion-Convolutional Neural Networks , 2015, NIPS.

[45]  Regina Barzilay,et al.  Deriving Neural Architectures from Sequence and Graph Kernels , 2017, ICML.

[46]  Alán Aspuru-Guzik,et al.  Convolutional Networks on Graphs for Learning Molecular Fingerprints , 2015, NIPS.

[47]  Max Welling,et al.  Semi-Supervised Classification with Graph Convolutional Networks , 2016, ICLR.

[48]  Joseph Gomes,et al.  MoleculeNet: a benchmark for molecular machine learning† †Electronic supplementary information (ESI) available. See DOI: 10.1039/c7sc02664a , 2017, Chemical science.

[49]  Jure Leskovec,et al.  Inductive Representation Learning on Large Graphs , 2017, NIPS.

[50]  Pietro Liò,et al.  Graph Attention Networks , 2017, ICLR.

[51]  Jan Eric Lenssen,et al.  Fast Graph Representation Learning with PyTorch Geometric , 2019, ArXiv.

[52]  Stephan Günnemann,et al.  Dual-Primal Graph Convolutional Networks , 2018, ArXiv.

[53]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[54]  Richard S. Zemel,et al.  Gated Graph Sequence Neural Networks , 2015, ICLR.

[55]  László Babai,et al.  Graph isomorphism in quasipolynomial time [extended abstract] , 2015, STOC.