Topology preserving maps as aggregations for Graph Convolutional Neural Networks

In Graph Convolutional Neural Networks, the capability of learning the representation of graph nodes comes at hand when dealing with graph analysis tasks, such as predicting node properties. Furthermore, node-level representations can be aggregated to obtain a single graph-level representation and predictor. This work explores an alternative route for defining the aggregation function compared to existing approaches. We propose a graph aggregator that exploits Generative Topographic Mapping (GTM) to transform a set of node-level representations into a single graph-level one. The integration of GTM in a GCNN pipeline allows to estimate node representation probability densities and projects them in a low-dimensional space, while retaining the information about their mutual similarity and topology. A novel dedicated training procedure is specifically designed to learn from these reduced representations instead of the complete initial data. Experimental results on several graph classification datasets show that this approach achieves competitive predictive performances with respect to the commonly adopted aggregation architectures present in the literature while holding a well-grounded theoretical framework.

[1]  Tongliang Liu,et al.  Graph Pooling for Graph Neural Networks: Progress, Challenges, and Opportunities , 2022, IJCAI.

[2]  Alessandro Sperduti,et al.  SOM-based aggregation for graph convolutional neural networks , 2020, Neural Comput. Appl..

[3]  A. Micheli,et al.  A Fair Comparison of Graph Neural Networks for Graph Classification , 2019, ICLR.

[4]  Natalia Gimelshein,et al.  PyTorch: An Imperative Style, High-Performance Deep Learning Library , 2019, NeurIPS.

[5]  Alessandro Sperduti,et al.  Universal Readout for Graph Convolutional Neural Networks , 2019, 2019 International Joint Conference on Neural Networks (IJCNN).

[6]  Yizhou Sun,et al.  Are Powerful Graph Neural Nets Necessary? A Dissection on Graph Classification , 2019, ArXiv.

[7]  Jure Leskovec,et al.  How Powerful are Graph Neural Networks? , 2018, ICLR.

[8]  Jure Leskovec,et al.  Hierarchical Graph Representation Learning with Differentiable Pooling , 2018, NeurIPS.

[9]  Yixin Chen,et al.  An End-to-End Deep Learning Architecture for Graph Classification , 2018, AAAI.

[10]  Frank Hutter,et al.  Decoupled Weight Decay Regularization , 2017, ICLR.

[11]  Jure Leskovec,et al.  Inductive Representation Learning on Large Graphs , 2017, NIPS.

[12]  Max Welling,et al.  Semi-Supervised Classification with Graph Convolutional Networks , 2016, ICLR.

[13]  Xavier Bresson,et al.  Convolutional Neural Networks on Graphs with Fast Localized Spectral Filtering , 2016, NIPS.

[14]  Mathias Niepert,et al.  Learning Convolutional Neural Networks for Graphs , 2016, ICML.

[15]  T. Guillot,et al.  SOPHIE velocimetry of Kepler transit candidates XVII. The physical properties of giant exoplanets within 400 days of period , 2015, 1511.00643.

[16]  Kurt Mehlhorn,et al.  Weisfeiler-Lehman Graph Kernels , 2011, J. Mach. Learn. Res..

[17]  Alessio Micheli,et al.  Neural Network for Graphs: A Contextual Constructive Approach , 2009, IEEE Transactions on Neural Networks.

[18]  Christopher M. Bishop,et al.  Developments of the generative topographic mapping , 1998, Neurocomputing.

[19]  Alessandro Sperduti,et al.  Supervised neural networks for the classification of structures , 1997, IEEE Trans. Neural Networks.

[20]  Jooyoung Park,et al.  Universal Approximation Using Radial-Basis-Function Networks , 1991, Neural Computation.

[21]  D. Rubin,et al.  Maximum likelihood from incomplete data via the EM - algorithm plus discussions on the paper , 1977 .

[22]  Oriol Vinyals,et al.  Order Matters: Sequence to sequence for sets , 2016, ICLR 2016.

[23]  Teuvo Kohonen,et al.  Self-organized formation of topologically correct feature maps , 2004, Biological Cybernetics.

[24]  Christopher M. Bishop,et al.  GTM: The Generative Topographic Mapping , 1998, Neural Computation.