Towards an understanding of benign overfitting in neural networks

Z Li, ZH Zhou, A Gretton - arXiv preprint arXiv:2106.03212, 2021 - arxiv.org
Modern machine learning models often employ a huge number of parameters and are
typically optimized to have zero training loss; yet surprisingly, they possess near-optimal …

Decoupling the depth and scope of graph neural networks

H Zeng, M Zhang, Y Xia, A Srivastava… - Advances in …, 2021 - proceedings.neurips.cc
State-of-the-art Graph Neural Networks (GNNs) have limited scalability with respect to the
graph and model sizes. On large graphs, increasing the model depth often means …

Why do larger models generalize better? A theoretical perspective via the XOR problem

A Brutzkus, A Globerson - International Conference on …, 2019 - proceedings.mlr.press
Empirical evidence suggests that neural networks with ReLU activations generalize better
with over-parameterization. However, there is currently no theoretical analysis that explains …

Rewiring with positional encodings for graph neural networks

R Brüel-Gabrielsson, M Yurochkin… - arXiv preprint arXiv …, 2022 - arxiv.org
Several recent works use positional encodings to extend the receptive fields of graph neural
network (GNN) layers equipped with attention mechanisms. These techniques, however …

Tree mover's distance: Bridging graph metrics and stability of graph neural networks

CY Chuang, S Jegelka - Advances in Neural Information …, 2022 - proceedings.neurips.cc
Understanding generalization and robustness of machine learning models fundamentally
relies on assuming an appropriate metric on the data space. Identifying such a metric is …

Sizeshiftreg: a regularization method for improving size-generalization in graph neural networks

D Buffelli, P Liò, F Vandin - Advances in Neural Information …, 2022 - proceedings.neurips.cc
In the past few years, graph neural networks (GNNs) have become the de facto model of
choice for graph classification. While, from the theoretical viewpoint, most GNNs can operate …

Interpreting graph neural networks for NLP with differentiable edge masking

MS Schlichtkrull, N De Cao, I Titov - arXiv preprint arXiv:2010.00577, 2020 - arxiv.org
Graph neural networks (GNNs) have become a popular approach to integrating structural
inductive biases into NLP models. However, there has been little work on interpreting them …

On provable benefits of depth in training graph convolutional networks

W Cong, M Ramezani… - Advances in Neural …, 2021 - proceedings.neurips.cc
Abstract Graph Convolutional Networks (GCNs) are known to suffer from performance
degradation as the number of layers increases, which is usually attributed to over …

Edge contraction pooling for graph neural networks

F Diehl - arXiv preprint arXiv:1905.10990, 2019 - arxiv.org
Graph Neural Network (GNN) research has concentrated on improving convolutional layers,
with little attention paid to developing graph pooling layers. Yet pooling layers can enable …

Graphnorm: A principled approach to accelerating graph neural network training

T Cai, S Luo, K Xu, D He, T Liu… - … Conference on Machine …, 2021 - proceedings.mlr.press
Normalization is known to help the optimization of deep neural networks. Curiously, different
architectures require specialized normalization methods. In this paper, we study what …