Turnitin
降AI改写
早检测系统
早降重系统
Turnitin-UK版
万方检测-期刊版
维普编辑部版
Grammarly检测
Paperpass检测
checkpass检测
PaperYY检测
Adversarial training for graph neural networks: Pitfalls, solutions, and new directions
L Gosch, S Geisler, D Sturm… - Advances in neural …, 2023 - proceedings.neurips.cc
Despite its success in the image domain, adversarial training did not (yet) stand out as an
effective defense for Graph Neural Networks (GNNs) against graph structure perturbations …
effective defense for Graph Neural Networks (GNNs) against graph structure perturbations …
Optimality of message-passing architectures for sparse graphs
We study the node classification problem on feature-decorated graphs in the sparse setting,
ie, when the expected degree of a node is $ O (1) $ in the number of nodes, in the fixed …
ie, when the expected degree of a node is $ O (1) $ in the number of nodes, in the fixed …
Bounding the expected robustness of graph neural networks subject to node feature attacks
Graph Neural Networks (GNNs) have demonstrated state-of-the-art performance in various
graph representation learning tasks. Recently, studies revealed their vulnerability to …
graph representation learning tasks. Recently, studies revealed their vulnerability to …
Hierarchical randomized smoothing
Y Scholten, J Schuchardt… - Advances in …, 2023 - proceedings.neurips.cc
Real-world data is complex and often consists of objects that can be decomposed into
multiple entities (eg images into pixels, graphs into interconnected nodes). Randomized …
multiple entities (eg images into pixels, graphs into interconnected nodes). Randomized …
Can Large Language Models Improve the Adversarial Robustness of Graph Neural Networks?
Graph neural networks (GNNs) are vulnerable to adversarial perturbations, especially for
topology attacks, and many methods that improve the robustness of GNNs have received …
topology attacks, and many methods that improve the robustness of GNNs have received …
Provable adversarial robustness for group equivariant tasks: Graphs, point clouds, molecules, and more
J Schuchardt, Y Scholten… - Advances in Neural …, 2023 - proceedings.neurips.cc
A machine learning model is traditionally considered robust if its prediction remains (almost)
constant under input perturbations with small norm. However, real-world tasks like molecular …
constant under input perturbations with small norm. However, real-world tasks like molecular …
Collaboration! Towards Robust Neural Methods for Routing Problems
Despite enjoying desirable efficiency and reduced reliance on domain expertise, existing
neural methods for vehicle routing problems (VRPs) suffer from severe robustness issues …
neural methods for vehicle routing problems (VRPs) suffer from severe robustness issues …
On the adversarial robustness of graph contrastive learning methods
Contrastive learning (CL) has emerged as a powerful framework for learning
representations of images and text in a self-supervised manner while enhancing model …
representations of images and text in a self-supervised manner while enhancing model …
Provable Robustness of (Graph) Neural Networks Against Data Poisoning and Backdoor Attacks
Generalization of machine learning models can be severely compromised by data
poisoning, where adversarial changes are applied to the training data. This vulnerability has …
poisoning, where adversarial changes are applied to the training data. This vulnerability has …
Boosting the adversarial robustness of graph neural networks: An ood perspective
Current defenses against graph attacks often rely on certain properties to eliminate structural
perturbations by identifying adversarial edges from normal edges. However, this …
perturbations by identifying adversarial edges from normal edges. However, this …