arXiv Analytics

Sign in

arXiv:2006.05205 [cs.LG]AbstractReferencesReviewsResources

On the Bottleneck of Graph Neural Networks and its Practical Implications

Uri Alon, Eran Yahav

Published 2020-06-09Version 1

Graph neural networks (GNNs) were shown to effectively learn from highly structured data containing elements (nodes) with relationships (edges) between them. GNN variants differ in how each node in the graph absorbs the information flowing from its neighbor nodes. In this paper, we highlight an inherent problem in GNNs: the mechanism of propagating information between neighbors creates a bottleneck when every node aggregates messages from its neighbors. This bottleneck causes the over-squashing of exponentially-growing information into fixed-size vectors. As a result, the graph fails to propagate messages flowing from distant nodes and performs poorly when the prediction task depends on long-range information. We demonstrate that the bottleneck hinders popular GNNs from fitting the training data. We show that GNNs that absorb incoming edges equally, like GCN and GIN, are more susceptible to over-squashing than other GNN types. We further show that existing, extensively-tuned, GNN-based models suffer from over-squashing and that breaking the bottleneck improves state-of-the-art results without any hyperparameter tuning or additional weights.

Related articles: Most relevant | Search more
arXiv:1905.04497 [cs.LG] (Published 2019-05-11)
Stability Properties of Graph Neural Networks
arXiv:1910.12241 [cs.LG] (Published 2019-10-27)
Pre-train and Learn: Preserve Global Information for Graph Neural Networks
arXiv:2006.06830 [cs.LG] (Published 2020-06-11)
Data Augmentation for Graph Neural Networks