论文标题
让Graph成为GO板:通过增强学习的图形神经网络的无梯度节点注射攻击
Let Graph be the Go Board: Gradient-free Node Injection Attack for Graph Neural Networks via Reinforcement Learning
论文作者
论文摘要
多年来,图形神经网络(GNN)引起了极大的关注,并广泛应用于需要稳定鲁棒性或强大安全标准的基本应用,例如产品建议和用户行为建模。在这些情况下,利用GNN的脆弱性并进一步降低其表现成为对对手的极其激励。以前的攻击者主要关注对现有图的结构扰动或节点注射,并由替代模型的梯度引导。尽管他们提供了令人鼓舞的结果,但仍然存在一些局限性。对于结构性扰动攻击,要发起拟议的攻击,对手需要操纵现有的图形拓扑,这在大多数情况下是不切实际的。尽管对于节点注射攻击,尽管更实用,但当前的方法需要培训替代模型来模拟白色框设置,这会导致替代体系结构与实际受害者模型的分歧时会大大降级。为了弥合这些差距,在本文中,我们研究了黑盒节点注射攻击的问题,而无需训练潜在的误导性替代模型。具体而言,我们将节点注射攻击模型为马尔可夫的决策过程,并提出了无梯度的优势演员评论家,即G2A2C,即以优势演员评论家的方式加强学习框架。通过直接查询受害者模型,G2A2C学会了以极有限的攻击预算注入高度恶意的节点,同时保持相似的节点特征分布。通过我们超过八个具有不同特征的基准数据集的全面实验,我们证明了我们提出的G2A2C的优越性能,而不是现有的最新攻击者。源代码可公开,网址为:https://github.com/jumxglhf/g2a2c}。
Graph Neural Networks (GNNs) have drawn significant attentions over the years and been broadly applied to essential applications requiring solid robustness or vigorous security standards, such as product recommendation and user behavior modeling. Under these scenarios, exploiting GNN's vulnerabilities and further downgrading its performance become extremely incentive for adversaries. Previous attackers mainly focus on structural perturbations or node injections to the existing graphs, guided by gradients from the surrogate models. Although they deliver promising results, several limitations still exist. For the structural perturbation attack, to launch a proposed attack, adversaries need to manipulate the existing graph topology, which is impractical in most circumstances. Whereas for the node injection attack, though being more practical, current approaches require training surrogate models to simulate a white-box setting, which results in significant performance downgrade when the surrogate architecture diverges from the actual victim model. To bridge these gaps, in this paper, we study the problem of black-box node injection attack, without training a potentially misleading surrogate model. Specifically, we model the node injection attack as a Markov decision process and propose Gradient-free Graph Advantage Actor Critic, namely G2A2C, a reinforcement learning framework in the fashion of advantage actor critic. By directly querying the victim model, G2A2C learns to inject highly malicious nodes with extremely limited attacking budgets, while maintaining a similar node feature distribution. Through our comprehensive experiments over eight acknowledged benchmark datasets with different characteristics, we demonstrate the superior performance of our proposed G2A2C over the existing state-of-the-art attackers. Source code is publicly available at: https://github.com/jumxglhf/G2A2C}.