B Wu, X Yang, S Pan, X Yuan - Proceedings of the 2022 ACM on Asia …, 2022 - dl.acm.org
Machine learning models are shown to face a severe threat from Model Extraction Attacks, where a well-trained private model owned by a service provider can be stolen by an attacker …
M Lin, T Xiao, E Dai, X Zhang… - Advances in Neural …, 2024 - proceedings.neurips.cc
Abstract Graph Contrastive Learning (GCL) has emerged as a popular unsupervised graph representation learning method. However, it has been shown that GCL is vulnerable to …
J Mu, B Wang, Q Li, K Sun, M Xu, Z Liu - Proceedings of the 2021 ACM …, 2021 - dl.acm.org
Graph Neural Networks (GNNs) have achieved state-of-the-art performance in various graph structure related tasks such as node classification and graph classification. However, GNNs …
B Wang, M Pang, Y Dong - … of the IEEE/CVF Conference on …, 2023 - openaccess.thecvf.com
Graph neural networks (GNNs) have achieved state-of-the-art performance in many graph- related tasks such as node classification. However, recent studies show that GNNs are …
Graph neural networks, a popular class of models effective in a wide range of graph-based learning tasks, have been shown to be vulnerable to adversarial attacks. While the majority …
Influence function, a method from robust statistics, measures the changes of model parameters or some functions about model parameters concerning the removal or …
L Lin, E Blaser, H Wang - Proceedings of the 28th ACM SIGKDD …, 2022 - dl.acm.org
Graph Convolutional Networks (GCNs) have fueled a surge of research interest due to their encouraging performance on graph learning tasks, but they are also shown vulnerability to …
Graph Neural Networks (GNNs) have emerged as a critical tool for optimizing and managing the complexities of the Internet of Things (IoT) in next-generation networks. This survey …
Recent studies indicate that Graph Neural Networks (GNNs) are vulnerable to adversarial attacks. Particularly, adversarially perturbing the graph structure, eg, flipping edges, can …