Message Passing Neural Networks (MPNNs) are instances of Graph Neural Networks that leverage the graph to send messages over the edges. This inductive bias leads to a phenomenon known as over-squashing, where a node feature is insensitive to information contained at distant nodes. Despite recent methods introduced to mitigate this issue, an understanding of the causes for oversquashing and of possible solutions are lacking. In this theoretical work, we prove that: (i) Neural network width can mitigate over-squashing, but at the cost of making the whole network more sensitive; (ii) Conversely, depth cannot help mitigate over-squashing: increasing the number of layers leads to over-squashing being dominated by vanishing gradients; (iii) The graph topology plays the greatest role, since over-squashing occurs between nodes at high commute time. Our analysis provides a unified framework to study different recent methods introduced to cope with over-squashing and serves as a justification for a class of methods that fall under graph rewiring.

On Over-Squashing in Message Passing Neural Networks: The Impact of Width, Depth, and Topology / Di Giovanni, F.; Giusti, L.; Barbero, F.; Luise, G.; Lio, P.; Bronstein, M.. - 202:(2023), pp. 7865-7885. (Intervento presentato al convegno International Conference on Machine Learning tenutosi a Honolulu).

On Over-Squashing in Message Passing Neural Networks: The Impact of Width, Depth, and Topology

Lio P.;
2023

Abstract

Message Passing Neural Networks (MPNNs) are instances of Graph Neural Networks that leverage the graph to send messages over the edges. This inductive bias leads to a phenomenon known as over-squashing, where a node feature is insensitive to information contained at distant nodes. Despite recent methods introduced to mitigate this issue, an understanding of the causes for oversquashing and of possible solutions are lacking. In this theoretical work, we prove that: (i) Neural network width can mitigate over-squashing, but at the cost of making the whole network more sensitive; (ii) Conversely, depth cannot help mitigate over-squashing: increasing the number of layers leads to over-squashing being dominated by vanishing gradients; (iii) The graph topology plays the greatest role, since over-squashing occurs between nodes at high commute time. Our analysis provides a unified framework to study different recent methods introduced to cope with over-squashing and serves as a justification for a class of methods that fall under graph rewiring.
2023
International Conference on Machine Learning
Graph neural networks; Machine learning; Multilayer neural networks; Topology
04 Pubblicazione in atti di convegno::04b Atto di convegno in volume
On Over-Squashing in Message Passing Neural Networks: The Impact of Width, Depth, and Topology / Di Giovanni, F.; Giusti, L.; Barbero, F.; Luise, G.; Lio, P.; Bronstein, M.. - 202:(2023), pp. 7865-7885. (Intervento presentato al convegno International Conference on Machine Learning tenutosi a Honolulu).
File allegati a questo prodotto
File Dimensione Formato  
DiGiovanni_On-Over-Squashing_2023.pdf

accesso aperto

Note: https://proceedings.mlr.press/v202/di-giovanni23a/di-giovanni23a.pdf
Tipologia: Versione editoriale (versione pubblicata con il layout dell'editore)
Licenza: Creative commons
Dimensione 2.29 MB
Formato Adobe PDF
2.29 MB Adobe PDF

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11573/1721265
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 21
  • ???jsp.display-item.citation.isi??? 0
social impact