Graph neural networks (GNNs) are commonly described as being permutation equivariant with respect to node relabeling in the graph. This symmetry of GNNs is often compared to …
Recently, many works studied the expressive power of graph neural networks (GNNs) by linking it to the $1 $-dimensional Weisfeiler-Leman algorithm ($1\text {-}\mathsf {WL} $) …
Graph neural networks are widely used tools for graph prediction tasks. Motivated by their empirical performance, prior works have developed generalization bounds for graph neural …
Inspired by their powerful representation ability on graph-structured data, Graph Convolution Networks (GCNs) have been widely applied to recommender systems, and have shown …
Graph Transformers, which incorporate self-attention and positional encoding, have recently emerged as a powerful architecture for various graph learning tasks. Despite their …
L Pasa, N Navarin, W Erb… - IEEE Transactions on …, 2023 - ieeexplore.ieee.org
Many neural networks for graphs are based on the graph convolution (GC) operator, proposed more than a decade ago. Since then, many alternative definitions have been …
The Weisfeiler-Leman algorithm ($1 $-WL) is a well-studied heuristic for the graph isomorphism problem. Recently, the algorithm has played a prominent role in understanding …
H Tang, Y Liu - arXiv preprint arXiv:2311.04561, 2023 - arxiv.org
In this paper, we develop data-dependent and algorithm-dependent generalization bounds for transductive learning algorithms in the context of information theory for the first time. We …
The fundamental principle of Graph Neural Networks (GNNs) is to exploit the structural information of the data by aggregating the neighboring nodes using agraph convolution' …