论文标题
基于偏见的通用对抗补丁攻击自动退房
Bias-based Universal Adversarial Patch Attack for Automatic Check-out
论文作者
论文摘要
对抗性示例是具有不可察觉的扰动的输入,这些输入很容易误导深层神经网络(DNNS)。最近,对抗性贴片,噪音仅限于一个小小的局部贴片,在现实情况下的可行性方面出现了。但是,现有的策略未能产生具有强大概括能力的对抗斑块。换句话说,对抗贴剂是特定于输入的,并且未能攻击所有班级的图像,尤其是在训练过程中看不见的图像。为了解决这个问题,本文提出了一个基于偏见的框架,以生成具有强大泛化能力的类无形通用对抗贴片,从而利用了模型的感知和语义偏见。关于感知偏见,由于DNN对纹理有很大的偏见,因此我们利用了艰难的例子,这些例子传达了强大的模型不确定性,并通过采用样式相似性来从中提取质地。补丁的先验更接近决策边界,并将促进攻击。为了进一步缓解对训练通用攻击中大量数据的大大依赖,我们进一步利用语义偏见。作为班级偏好,通过最大化多级别的利润来帮助普遍培训来引入和追求原型。在数字世界(RPC,最大的ACO相关数据集)和物理世界情景(Taobao和JD,世界上最大的在线购物平台)中进行了自动检查(ACO)作为典型情况,包括数字世界(RPC,最大的ACO与ACO相关数据集)的广泛实验。实验结果表明,我们提出的框架优于最先进的对抗斑块攻击方法。
Adversarial examples are inputs with imperceptible perturbations that easily misleading deep neural networks(DNNs). Recently, adversarial patch, with noise confined to a small and localized patch, has emerged for its easy feasibility in real-world scenarios. However, existing strategies failed to generate adversarial patches with strong generalization ability. In other words, the adversarial patches were input-specific and failed to attack images from all classes, especially unseen ones during training. To address the problem, this paper proposes a bias-based framework to generate class-agnostic universal adversarial patches with strong generalization ability, which exploits both the perceptual and semantic bias of models. Regarding the perceptual bias, since DNNs are strongly biased towards textures, we exploit the hard examples which convey strong model uncertainties and extract a textural patch prior from them by adopting the style similarities. The patch prior is more close to decision boundaries and would promote attacks. To further alleviate the heavy dependency on large amounts of data in training universal attacks, we further exploit the semantic bias. As the class-wise preference, prototypes are introduced and pursued by maximizing the multi-class margin to help universal training. Taking AutomaticCheck-out (ACO) as the typical scenario, extensive experiments including white-box and black-box settings in both digital-world(RPC, the largest ACO related dataset) and physical-world scenario(Taobao and JD, the world' s largest online shopping platforms) are conducted. Experimental results demonstrate that our proposed framework outperforms state-of-the-art adversarial patch attack methods.