Depth Scaling in Graph Neural Networks: Understanding the Flat Curve Behavior

Onderzoeksoutput: Articlepeer review


Training deep Graph Neural Networks (GNNs) has proved to be a challenging task. A key goal of many new GNN architectures is to enable the depth scaling seen in other types of deep learning models. However, unlike deep learning methods in other domains, deep GNNs do not show significant performance boosts when compared to their shallow counterparts (resulting in a flat curve of performance over depth). In this work, we investigate some of the reasons why this goal of depth still eludes GNN researchers. We also question the effectiveness of current methods to train deep GNNs and show evidence of different types of pathological behavior in these networks. Our results suggest that current approaches hide the problems with deep GNNs rather than solve them, as current deep GNNs are only as discriminative as their respective shallow versions.
Originele taal-2English
Aantal pagina's22
TijdschriftTransactions on Machine Learning Research (TMLR)
StatusPublished - 2024


Duik in de onderzoeksthema's van 'Depth Scaling in Graph Neural Networks: Understanding the Flat Curve Behavior'. Samen vormen ze een unieke vingerafdruk.

Citeer dit