W Jiang, H Liu, H Xiong - arXiv preprint arXiv:2312.12477, 2023 - arxiv.org
Graph Neural Networks (GNNs) have emerged as powerful representation learning tools for capturing complex dependencies within diverse graph-structured data. Despite their …
G Chen, Y Qian, B Wang, L Li - arXiv preprint arXiv:2310.18167, 2023 - arxiv.org
The large language models have achieved superior performance on various natural language tasks. One major drawback of such approaches is they are resource-intensive in …
Z Wang, B Xu, Y Yuan, H Shen, X Cheng - Proceedings of the 47th …, 2024 - dl.acm.org
Graph contrastive learning (GCL), standing as the dominant paradigm in the realm of graph pre-training, has yielded considerable progress. Nonetheless, its capacity for out-of …
H Yang, X Pei, K Yuan - arXiv preprint arXiv:2406.00764, 2024 - arxiv.org
Due to the performance degradation of graph neural networks (GNNs) under distribution shifts, the work on out-of-distribution (OOD) generalization on graphs has received …