论文标题
探索面部伪造检测的频率对抗攻击
Exploring Frequency Adversarial Attacks for Face Forgery Detection
论文作者
论文摘要
各种面部操纵技术引起了公众在道德,安全和隐私方面的关注。尽管现有的面部伪造分类器在检测假图像方面实现了有希望的性能,但这些方法容易受到对抗性示例的影响,并且在像素上注入了不可察觉的扰动。同时,许多面对伪造的探测器总是将真实面孔和假面之间的频率多样性作为关键线索。在本文中,我们提出了一种针对面部伪造探测器的频率对抗攻击方法,而不是将对抗性扰动注入空间域。具体而言,我们在输入图像上应用离散的余弦变换(DCT),并引入一个融合模块,以捕获频域中对手的显着区域。与空间域中现有的对抗攻击(例如FGSM,PGD)相比,我们的方法对人类观察者来说更不可感知,并且不会降低原始图像的视觉质量。此外,受到元学习概念的启发,我们还提出了一种在空间和频域中进行攻击的混合对抗攻击。广泛的实验表明,所提出的方法不仅愚弄了基于空间的检测器,而且还可以有效地基于最新的频率检测器。此外,提出的频率攻击可增强面部伪造探测器的可转移性,作为黑盒攻击。
Various facial manipulation techniques have drawn serious public concerns in morality, security, and privacy. Although existing face forgery classifiers achieve promising performance on detecting fake images, these methods are vulnerable to adversarial examples with injected imperceptible perturbations on the pixels. Meanwhile, many face forgery detectors always utilize the frequency diversity between real and fake faces as a crucial clue. In this paper, instead of injecting adversarial perturbations into the spatial domain, we propose a frequency adversarial attack method against face forgery detectors. Concretely, we apply discrete cosine transform (DCT) on the input images and introduce a fusion module to capture the salient region of adversary in the frequency domain. Compared with existing adversarial attacks (e.g. FGSM, PGD) in the spatial domain, our method is more imperceptible to human observers and does not degrade the visual quality of the original images. Moreover, inspired by the idea of meta-learning, we also propose a hybrid adversarial attack that performs attacks in both the spatial and frequency domains. Extensive experiments indicate that the proposed method fools not only the spatial-based detectors but also the state-of-the-art frequency-based detectors effectively. In addition, the proposed frequency attack enhances the transferability across face forgery detectors as black-box attacks.