Many efforts have been paid to enhance Graph Convolutional Network from the perspective of propagation under the philosophy that “Propagation is the essence of the GCNNs”. Unfortunately, its adverse effect is over-smoothing, which makes the performance dramatically drop. To prevent the over-smoothing, many variants are presented. However, the perspective of propagation can’t provide an intuitive and unified interpretation to their effect on prevent over-smoothing. In this paper, we aim at providing a novel explanation to the question of “Why do attributes propagate in GCNNs?”. which not only gives the essence of the oversmoothing, but also illustrates why the GCN extensions, including multi-scale GCN and GCN with initial residual, can improve the performance. To this end, an intuitive Graph Representation Learning (GRL) framework is presented. GRL simply constrains the node representation similar with the original attribute, and encourages the connected nodes possess similar representations (pairwise constraint). Based on the proposed GRL, exiting GCN and its extensions can be proved as different numerical optimization algorithms, such as gradient descent, of our proposed GRL framework. Inspired by the superiority of conjugate gradient descent compared to common gradient descent, a novel Graph Conjugate Convolutional (GCC) network is presented to approximate the solution to GRL with fast convergence. Specifically, GCC adopts the obtained information of the last layer, which can be represented as the difference between the input and output of the last layer, as the input to the next layer. Extensive experiments demonstrate the superior performance of GCC. Introduction Graph Neural Networks (GNNs) (Wu et al. 2020; Xu et al. 2019) have become a hot topic in deep learning for their potentials in modeling irregular data. GNNs have been widely used and achieved state-of-the-art performance in many fields, such as computer vision, natural language processing (Yang et al. 2020), traffic forecasting, chemistry and medical analysis, etc. Existing GNNs fall into two categories, spectral methods (Defferrard, Bresson, and Vandergheynst 2016) and spatial ones (Hamilton, Ying, and Leskovec 2017; Gilmer et al. 2017; Yang et al. 2019b,a; Jin et al. 2019, 2020, 2021). ∗Corresponding author. Copyright c © 2021, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved. Graph Convolutional Network (GCN) (Kipf and Welling 2017), which is a simple, well-behaved and insightful GNN, bridges above two perspectives by proving that the propagation can be motivated from a first-order approximation of spectral graph convolutions. Recently progress also demonstrates the equivalent of spatial and spectral ones (Balcilar et al. 2020). Many efforts have been paid to enhance GCN from the perspective of propagation (Gilmer et al. 2017), such as learnable propagation weights in Graph Attention Network (GAT) (Velickovic et al. 2018), Gated Attention Network (GaAN) (Zhang et al. 2018) and Probabilistic GCN (Yang et al. 2020), structural neighbourhood in Geom-GCN (Pei et al. 2020) and multi-scale (multi-hop) combination in N-GCN (Abu-El-Haija et al. 2019a), MixHop (Abu-El-Haija et al. 2019b), LanczosNet (Liao et al. 2019) and Krylov GCN (Luan et al. 2019). The common philosophy of them is: “Propagation is the essence of the GCNNs”. And, the success of GCNs attributes to the Laplacian smoothing induced by the propagation (Li, Han, and Wu 2018). Unfortunately, the most serious issue of GNNs is the over-smoothing, which makes the performance dramatically drop, caused by the multiple propagations via stacking multiple graph convolution layers. Recently, (Oono and Suzuki 2020) shows the the exponential loss of expressive power of GNNs by generalizing the forward propagation of a GCN as a specific dynamical system. To prevent over-smoothing, two kinds of methods are proposed. On one hand, methods in the first category constrain the propagation. Disentangled GCN (Ma et al. 2019) makes each attribute only be propagated on part of the edges. DropEdge (Rong et al. 2020) randomly removes a certain number of edges from the input graph at each training epoch to reduce the adverse effect of message passing. On the other hand, methods in the second category constrain the propagation result with the original attributes. PageRank-GCN (Klicpera, Bojchevski, and Günnemann 2019) integrates personalized PageRank to GCN to combine the original attribute. JKNet (Xu et al. 2018) employs dense connections for multi-hop message passing, while DeepGCN (Li et al. 2019) and (GCNII) (Chen et al. 2020) incorporates residual layers into GCNs to facilitate the development of deep architectures. However, the perspective of propagation can’t provide an intuitive and unified interpretation to their effect on preventing over-smoothing. The Thirty-Fifth AAAI Conference on Artificial Intelligence (AAAI-21)
[1]
Lise Getoor,et al.
Collective Classification in Network Data
,
2008,
AI Mag..
[2]
Jian Sun,et al.
Deep Residual Learning for Image Recognition
,
2015,
2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[3]
Ziyang Liu,et al.
BiTe-GCN: A New GCN Architecture via BidirectionalConvolution of Topology and Features on Text-Rich Networks
,
2020,
ArXiv.
[4]
Max Welling,et al.
Semi-Supervised Classification with Graph Convolutional Networks
,
2016,
ICLR.
[5]
Ruslan Salakhutdinov,et al.
Revisiting Semi-Supervised Learning with Graph Embeddings
,
2016,
ICML.
[6]
Ken-ichi Kawarabayashi,et al.
Representation Learning on Graphs with Jumping Knowledge Networks
,
2018,
ICML.
[7]
Xiao-Ming Wu,et al.
Deeper Insights into Graph Convolutional Networks for Semi-Supervised Learning
,
2018,
AAAI.
[8]
Tingyang Xu,et al.
DropEdge: Towards Deep Graph Convolutional Networks on Node Classification
,
2020,
ICLR.
[9]
Gene H. Golub,et al.
Matrix computations (3rd ed.)
,
1996
.
[10]
Le Song,et al.
Stochastic Training of Graph Convolutional Networks with Variance Reduction
,
2017,
ICML.
[11]
Kilian Q. Weinberger,et al.
Simplifying Graph Convolutional Networks
,
2019,
ICML.
[12]
Taiji Suzuki,et al.
Graph Neural Networks Exponentially Lose Expressive Power for Node Classification
,
2019,
ICLR.
[13]
Henry P. Decell.
An application of the Cayley-Hamilton theorem to generalized matrix inversion.
,
1965
.
[14]
Stephan Günnemann,et al.
Predict then Propagate: Graph Neural Networks meet Personalized PageRank
,
2018,
ICLR.
[15]
Jimmy Ba,et al.
Adam: A Method for Stochastic Optimization
,
2014,
ICLR.
[16]
Sergey Ioffe,et al.
Rethinking the Inception Architecture for Computer Vision
,
2015,
2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[17]
Vladlen Koltun,et al.
Robust continuous clustering
,
2017,
Proceedings of the National Academy of Sciences.
[18]
Paul Honeine,et al.
Bridging the Gap Between Spectral and Spatial Domains in Graph Neural Networks
,
2020,
ArXiv.
[19]
Joonseok Lee,et al.
N-GCN: Multi-scale Graph Convolution for Semi-supervised Node Classification
,
2018,
UAI.
[20]
Liang Yang,et al.
Masked Graph Convolutional Network
,
2019,
IJCAI.
[21]
Kevin Chen-Chuan Chang,et al.
Geom-GCN: Geometric Graph Convolutional Networks
,
2020,
ICLR.
[22]
Weixiong Zhang,et al.
Graph Convolutional Networks Meet Markov Random Fields: Semi-Supervised Community Detection in Attribute Networks
,
2019,
AAAI.
[23]
Samuel S. Schoenholz,et al.
Neural Message Passing for Quantum Chemistry
,
2017,
ICML.
[24]
Jure Leskovec,et al.
Inductive Representation Learning on Large Graphs
,
2017,
NIPS.
[25]
Pietro Liò,et al.
Graph Attention Networks
,
2017,
ICLR.
[26]
Xiaochun Cao,et al.
Topology Optimization based Graph Convolutional Network
,
2019,
IJCAI.
[27]
Samy Bengio,et al.
Cluster-GCN: An Efficient Algorithm for Training Deep and Large Graph Convolutional Networks
,
2019,
KDD.
[28]
Kristina Lerman,et al.
MixHop: Higher-Order Graph Convolutional Architectures via Sparsified Neighborhood Mixing
,
2019,
ICML.
[29]
Xiaolong Li,et al.
GeniePath: Graph Neural Networks with Adaptive Receptive Paths
,
2018,
AAAI.
[30]
Yaliang Li,et al.
Simple and Deep Graph Convolutional Networks
,
2020,
ICML.
[31]
Hao Ma,et al.
GaAN: Gated Attention Networks for Learning on Large and Spatiotemporal Graphs
,
2018,
UAI.
[32]
Bernard Ghanem,et al.
DeepGCNs: Can GCNs Go As Deep As CNNs?
,
2019,
2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[33]
Philip S. Yu,et al.
GCN for HIN via Implicit Utilization of Attention and Meta-Paths
,
2020,
IEEE Transactions on Knowledge and Data Engineering.
[34]
Zhizhen Zhao,et al.
LanczosNet: Multi-Scale Deep Graph Convolutional Networks
,
2019,
ICLR.
[35]
Philip S. Yu,et al.
A Comprehensive Survey on Graph Neural Networks
,
2019,
IEEE Transactions on Neural Networks and Learning Systems.
[36]
Xavier Bresson,et al.
Convolutional Neural Networks on Graphs with Fast Localized Spectral Filtering
,
2016,
NIPS.
[37]
Doina Precup,et al.
Break the Ceiling: Stronger Multi-scale Deep Graph Convolutional Networks
,
2019,
NeurIPS.
[38]
Wenwu Zhu,et al.
Disentangled Graph Convolutional Networks
,
2019,
ICML.
[39]
Jure Leskovec,et al.
How Powerful are Graph Neural Networks?
,
2018,
ICLR.
[40]
Xiaochun Cao,et al.
Graph Attention Topic Modeling Network
,
2020,
WWW.
[41]
Stuart Geman,et al.
Statistical methods for tomographic image reconstruction
,
1987
.
[42]
Xiaochun Cao,et al.
Probabilistic Graph Convolutional Network via Topology-Constrained Latent Space Model
,
2020,
IEEE Transactions on Cybernetics.