Graph Explicit Neural Networks: Explicitly Encoding Graphs for Efficient and Accurate Inference

As the state-of-the-art graph learning models, the message passing based neural networks (MPNNs) implicitly use the graph topology as the "pathways" to propagate node features. This implicit use of graph topology induces the MPNNs' over-reliance on (node) features and high inference latency, which hinders their large-scale applications in industrial contexts. To mitigate these weaknesses, we propose the Graph Explicit Neural Network (GENN) framework. GENN can be flexibly applied to various MPNNs and improves them by providing more efficient and accurate inference that is robust in feature-constrained settings. Specifically, we carefully incorporate recent developments in network embedding methods to efficiently prioritize the graph topology for inference. From this vantage, GENN explicitly encodes the topology as an important source of information to mitigate the reliance on node features. Moreover, by adopting knowledge distillation (KD) techniques, GENN takes an MPNN as the teacher to supervise the training for better effectiveness while avoiding the teacher's high inference latency. Empirical results show that our GENN infers dramatically faster than its MPNN teacher by 40x-78x. In terms of accuracy, GENN yields significant gains (more than 40%) for its MPNN teacher when the node features are limited based on our explicit encoding. Moreover, GENN outperforms the MPNN teacher even in feature-rich settings thanks to our KD design.

[1]  Gian Luca Scoccia,et al.  An empirical study of privacy labels on the Apple iOS mobile app store , 2022, 2022 IEEE/ACM 9th International Conference on Mobile Software Engineering and Systems (MobileSoft).

[2]  N. Shadbolt,et al.  Goodbye Tracking? Impact of iOS App Tracking Transparency and Privacy Labels , 2022, FAccT.

[3]  Bryan Perozzi,et al.  GraphWorld: Fake Graphs Bring Real Insights for GNNs , 2022, Knowledge Discovery and Data Mining.

[4]  Yizhou Sun,et al.  Graph-less Neural Networks: Teaching Old MLPs New Tricks via Distillation , 2021, ICLR.

[5]  Erjin Zhou,et al.  Graph-MLP: Node Classification without Message Passing in Graph , 2021, ArXiv.

[6]  Rajgopal Kannan,et al.  Accelerating Large Scale Real-Time GNN Inference using Channel Pruning , 2021, Proc. VLDB Endow..

[7]  Roger Wattenhofer,et al.  Should Graph Neural Networks Use Features, Edges, Or Both? , 2021, ArXiv.

[8]  Chuan Shi,et al.  Extract the Knowledge of Graph Neural Networks and Go Beyond it: An Effective Knowledge Distillation Framework , 2021, WWW.

[9]  Silvio Lattanzi,et al.  InstantEmbedding: Efficient Local Node Representations , 2020, ArXiv.

[10]  Pietro Lio,et al.  Learned Low Precision Graph Neural Networks , 2020, ArXiv.

[11]  Bryan Hooi,et al.  NodeAug: Semi-Supervised Node Classification with Data Augmentation , 2020, KDD.

[12]  Yaliang Li,et al.  Simple and Deep Graph Convolutional Networks , 2020, ICML.

[13]  Jianping Gou,et al.  Knowledge Distillation: A Survey , 2020, International Journal of Computer Vision.

[14]  Walid Krichene,et al.  Neural Collaborative Filtering vs. Matrix Factorization Revisited , 2020, RecSys.

[15]  J. Leskovec,et al.  Open Graph Benchmark: Datasets for Machine Learning on Graphs , 2020, NeurIPS.

[16]  Yuxiao Dong,et al.  Microsoft Academic Graph: When experts are not enough , 2020, Quantitative Science Studies.

[17]  Karl Aberer,et al.  On Node Features for Graph Neural Networks , 2019, ArXiv.

[18]  Yizhou Sun,et al.  Layer-Dependent Importance Sampling for Training Deep and Large Graph Convolutional Networks , 2019, NeurIPS.

[19]  Steven Skiena,et al.  Fast and Accurate Network Embeddings via Very Sparse Random Projection , 2019, CIKM.

[20]  Jure Leskovec,et al.  Position-aware Graph Neural Networks , 2019, ICML.

[21]  Samy Bengio,et al.  Cluster-GCN: An Efficient Algorithm for Training Deep and Large Graph Convolutional Networks , 2019, KDD.

[22]  Kilian Q. Weinberger,et al.  Simplifying Graph Convolutional Networks , 2019, ICML.

[23]  Chien Chou,et al.  Privacy behavior profiles of underage Facebook users , 2019, Comput. Educ..

[24]  Stephan Günnemann,et al.  Predict then Propagate: Graph Neural Networks meet Personalized PageRank , 2018, ICLR.

[25]  Zachary Chase Lipton,et al.  Born Again Neural Networks , 2018, ICML.

[26]  Xiao Wang,et al.  Billion-Scale Network Embedding with Iterative Random Projection , 2018, 2018 IEEE International Conference on Data Mining (ICDM).

[27]  Jian Li,et al.  Network Embedding as Matrix Factorization: Unifying DeepWalk, LINE, PTE, and node2vec , 2017, WSDM.

[28]  Stephan Günnemann,et al.  Deep Gaussian Embedding of Graphs: Unsupervised Inductive Learning via Ranking , 2017, ICLR.

[29]  Jure Leskovec,et al.  Inductive Representation Learning on Large Graphs , 2017, NIPS.

[30]  Qun Liu,et al.  Lexically Constrained Decoding for Sequence Generation Using Grid Beam Search , 2017, ACL.

[31]  Max Welling,et al.  Semi-Supervised Classification with Graph Convolutional Networks , 2016, ICLR.

[32]  Jure Leskovec,et al.  node2vec: Scalable Feature Learning for Networks , 2016, KDD.

[33]  Xavier Bresson,et al.  Convolutional Neural Networks on Graphs with Fast Localized Spectral Filtering , 2016, NIPS.

[34]  Natalie D. Enright Jerger,et al.  Proteus: Exploiting Numerical Precision Variability in Deep Neural Networks , 2016, ICS.

[35]  Ruslan Salakhutdinov,et al.  Revisiting Semi-Supervised Learning with Graph Embeddings , 2016, ICML.

[36]  Song Han,et al.  Learning both Weights and Connections for Efficient Neural Network , 2015, NIPS.

[37]  Pritish Narayanan,et al.  Deep Learning with Limited Numerical Precision , 2015, ICML.

[38]  Steven Skiena,et al.  DeepWalk: online learning of social representations , 2014, KDD.

[39]  Joan Bruna,et al.  Spectral Networks and Locally Connected Networks on Graphs , 2013, ICLR.

[40]  Sonia Livingstone,et al.  Risky Social Networking Practices Among "Underage" Users: Lessons for Evidence-Based Policy , 2013, J. Comput. Mediat. Commun..

[41]  Z. Yang Multi-layer Perceptron , 2010, Handbook of Machine Learning.

[42]  David J. DeWitt,et al.  Mondrian Multidimensional K-Anonymity , 2006, 22nd International Conference on Data Engineering (ICDE'06).

[43]  L. Sweeney,et al.  k-Anonymity: A Model for Protecting Privacy , 2002, Int. J. Uncertain. Fuzziness Knowl. Based Syst..

[44]  S T Roweis,et al.  Nonlinear dimensionality reduction by locally linear embedding. , 2000, Science.

[45]  J. Tenenbaum,et al.  A global geometric framework for nonlinear dimensionality reduction. , 2000, Science.

[46]  Kurt Hornik,et al.  Multilayer feedforward networks are universal approximators , 1989, Neural Networks.

[47]  OUP accepted manuscript , 2022, Journal of European Competition Law and Practice.

[48]  L. Gool,et al.  Supplementary Material for “Towards Efficient Graph Convolutional Networks for Point Cloud Handling” , 2021 .

[49]  Guigang Zhang,et al.  Deep Learning , 2016, Int. J. Semantic Comput..

[50]  James M. Joyce Kullback-Leibler Divergence , 2011, International Encyclopedia of Statistical Science.