论文标题
fake:通过文本到图像生成模型生成的假图像的检测和归因
DE-FAKE: Detection and Attribution of Fake Images Generated by Text-to-Image Generation Models
论文作者
论文摘要
基于及时描述生成图像的文本到图像生成模型在过去几个月中引起了越来越多的关注。尽管表现令人鼓舞,但这些模型引起了人们对滥用其产生的假图像的担忧。为了解决这个问题,我们开创了一项系统的研究,该研究对文本到图像生成模型产生的假图像的检测和归因。具体而言,我们首先构建了一个机器学习分类器,以检测各种文本到图像生成模型产生的假图像。然后,我们将这些假图像归因于他们的源模型,以便模型所有者可以对其模型的滥用负责。我们进一步研究了生成假图像的提示如何影响检测和归因。我们对四个流行的文本到图像生成模型进行了广泛的实验,包括dall $ \ cdot $ e 2,稳定的扩散,滑行和潜在扩散以及两个基准提示图数据集。经验结果表明,(1)各种模型生成的假图像可以与真实模型区分开,因为存在来自不同模型的假图像共享的常见人工制品; (2)假图像可以有效地归因于其源模型,因为不同的模型在生成的图像中留下了独特的指纹; (3)提示``人物''主题或25至75之间的长度使模型能够生成具有更高真实性的假图像。所有发现都有助于社区对文本到图像生成模型造成的威胁的见解。我们呼吁社区对同行解决方案(如我们的解决方案)的考虑,以反对快速发展的假图像产生。
Text-to-image generation models that generate images based on prompt descriptions have attracted an increasing amount of attention during the past few months. Despite their encouraging performance, these models raise concerns about the misuse of their generated fake images. To tackle this problem, we pioneer a systematic study on the detection and attribution of fake images generated by text-to-image generation models. Concretely, we first build a machine learning classifier to detect the fake images generated by various text-to-image generation models. We then attribute these fake images to their source models, such that model owners can be held responsible for their models' misuse. We further investigate how prompts that generate fake images affect detection and attribution. We conduct extensive experiments on four popular text-to-image generation models, including DALL$\cdot$E 2, Stable Diffusion, GLIDE, and Latent Diffusion, and two benchmark prompt-image datasets. Empirical results show that (1) fake images generated by various models can be distinguished from real ones, as there exists a common artifact shared by fake images from different models; (2) fake images can be effectively attributed to their source models, as different models leave unique fingerprints in their generated images; (3) prompts with the ``person'' topic or a length between 25 and 75 enable models to generate fake images with higher authenticity. All findings contribute to the community's insight into the threats caused by text-to-image generation models. We appeal to the community's consideration of the counterpart solutions, like ours, against the rapidly-evolving fake image generation.