论文标题
朝着对图神经网络的更实用的对抗性攻击
Towards More Practical Adversarial Attacks on Graph Neural Networks
论文作者
论文摘要
我们在新颖且现实的约束下研究了对图形神经网络(GNN)的黑盒攻击:攻击者只能访问网络中的一部分节点,他们只能攻击其中的少数。在此设置下,节点选择步骤至关重要。我们证明,GNN模型的结构归纳偏见可以成为这种攻击的有效来源。具体而言,通过利用GNN的向后传播与随机步行的向后传播之间的连接,我们表明,可以通过梯度和类似于Pagerank的重要性得分之间的连接将基于梯度的白色盒子攻击推广到黑框设置。在实践中,我们发现基于这种重要性得分的攻击确实会增加分类损失的幅度很大,但是它们无法显着提高错误分类率。我们的理论和经验分析表明,损失和错误分类率之间存在差异,因为当攻击节点的数量增加时,后者会呈现出返回的模式。因此,我们提出了一个贪婪的程序,以纠正考虑减少回报模式的重要性得分。实验结果表明,所提出的程序可以显着提高实际数据上常见GNN的错误分类率,而无需访问模型参数或预测。
We study the black-box attacks on graph neural networks (GNNs) under a novel and realistic constraint: attackers have access to only a subset of nodes in the network, and they can only attack a small number of them. A node selection step is essential under this setup. We demonstrate that the structural inductive biases of GNN models can be an effective source for this type of attacks. Specifically, by exploiting the connection between the backward propagation of GNNs and random walks, we show that the common gradient-based white-box attacks can be generalized to the black-box setting via the connection between the gradient and an importance score similar to PageRank. In practice, we find attacks based on this importance score indeed increase the classification loss by a large margin, but they fail to significantly increase the mis-classification rate. Our theoretical and empirical analyses suggest that there is a discrepancy between the loss and mis-classification rate, as the latter presents a diminishing-return pattern when the number of attacked nodes increases. Therefore, we propose a greedy procedure to correct the importance score that takes into account of the diminishing-return pattern. Experimental results show that the proposed procedure can significantly increase the mis-classification rate of common GNNs on real-world data without access to model parameters nor predictions.