How neural networks extrapolate: From feedforward to graph neural networks

K Xu, M Zhang, J Li, SS Du, K Kawarabayashi… - arXiv preprint arXiv …, 2020 - arxiv.org
We study how neural networks trained by gradient descent extrapolate, ie, what they learn
outside the support of the training distribution. Previous works report mixed empirical results …

Learning theory can (sometimes) explain generalisation in graph neural networks

P Esser, L Chennuru Vankadara… - Advances in Neural …, 2021 - proceedings.neurips.cc
In recent years, several results in the supervised learning setting suggested that classical
statistical learning-theoretic measures, such as VC dimension, do not adequately explain …

Optimization of graph neural networks: Implicit acceleration by skip connections and more depth

K Xu, M Zhang, S Jegelka… - … on Machine Learning, 2021 - proceedings.mlr.press
Abstract Graph Neural Networks (GNNs) have been studied through the lens of expressive
power and generalization. However, their optimization properties are less well understood …

On the bottleneck of graph neural networks and its practical implications

U Alon, E Yahav - arXiv preprint arXiv:2006.05205, 2020 - arxiv.org
Since the proposal of the graph neural network (GNN) by Gori et al.(2005) and Scarselli et
al.(2008), one of the major problems in training GNNs was their struggle to propagate …

What can linearized neural networks actually say about generalization?

G Ortiz-Jiménez… - Advances in Neural …, 2021 - proceedings.neurips.cc
For certain infinitely-wide neural networks, the neural tangent kernel (NTK) theory fully
characterizes generalization, but for the networks used in practice, the empirical NTK only …

Towards understanding generalization of graph neural networks

H Tang, Y Liu - International Conference on Machine …, 2023 - proceedings.mlr.press
Graph neural networks (GNNs) are widely used in machine learning for graph-structured
data. Even though GNNs have achieved remarkable success in real-world applications …

Graph neural networks for learning equivariant representations of neural networks

M Kofinas, B Knyazev, Y Zhang, Y Chen… - arXiv preprint arXiv …, 2024 - arxiv.org
Neural networks that process the parameters of other neural networks find applications in
domains as diverse as classifying implicit neural representations, generating neural network …

A non-asymptotic analysis of oversmoothing in graph neural networks

X Wu, Z Chen, W Wang, A Jadbabaie - arXiv preprint arXiv:2212.10701, 2022 - arxiv.org
Oversmoothing is a central challenge of building more powerful Graph Neural Networks
(GNNs). While previous works have only demonstrated that oversmoothing is inevitable …

Graph neural networks are inherently good generalizers: Insights by bridging gnns and mlps

C Yang, Q Wu, J Wang, J Yan - arXiv preprint arXiv:2212.09034, 2022 - arxiv.org
Graph neural networks (GNNs), as the de-facto model class for representation learning on
graphs, are built upon the multi-layer perceptrons (MLP) architecture with additional …

Graph neural networks inspired by classical iterative algorithms

Y Yang, T Liu, Y Wang, J Zhou, Q Gan… - International …, 2021 - proceedings.mlr.press
Despite the recent success of graph neural networks (GNN), common architectures often
exhibit significant limitations, including sensitivity to oversmoothing, long-range …