论文标题

优质的指导素描到光明图像综合

Quality Guided Sketch-to-Photo Image Synthesis

论文作者

Osahor, Uche, Kazemi, Hadi, Dabouei, Ali, Nasrabadi, Nasser

论文摘要

艺术家绘制的面部草图被广泛用于视觉识别应用程序,主要由执法机构使用,但是这些草图的质量取决于艺术家清楚地复制所有可以帮助捕获主题真正身份的所有关键面部特征的能力。最近的作品试图将这些草图合成为合理的视觉图像,以改善视觉识别和识别。但是,从草图中综合的照片现实图像被证明是一项更具挑战性的任务,尤其是对于诸如可疑识别之类的敏感应用程序。 In this work, we propose a novel approach that adopts a generative adversarial network that synthesizes a single sketch into multiple synthetic images with unique attributes like hair color, sex, etc. We incorporate a hybrid discriminator which performs attribute classification of multiple target attributes, a quality guided encoder that minimizes the perceptual dissimilarity of the latent space embedding of the synthesized and real image at different layers in the network and an在整个培训过程中保持综合图像的身份的身份保留网络。我们的方法旨在改善合成图像的视觉吸引力,同时将多个属性分配到发电机中,而不会损害合成图像的身份。我们使用XDOG过滤器为Celeba,WVU多模式和Celeba-HQ数据集合成草图,并从辅助生成器中培训了来自Cuhk,IIT-D和Feret数据集的草图。与当前的最新状态相比,我们的结果令人印象深刻。

Facial sketches drawn by artists are widely used for visual identification applications and mostly by law enforcement agencies, but the quality of these sketches depend on the ability of the artist to clearly replicate all the key facial features that could aid in capturing the true identity of a subject. Recent works have attempted to synthesize these sketches into plausible visual images to improve visual recognition and identification. However, synthesizing photo-realistic images from sketches proves to be an even more challenging task, especially for sensitive applications such as suspect identification. In this work, we propose a novel approach that adopts a generative adversarial network that synthesizes a single sketch into multiple synthetic images with unique attributes like hair color, sex, etc. We incorporate a hybrid discriminator which performs attribute classification of multiple target attributes, a quality guided encoder that minimizes the perceptual dissimilarity of the latent space embedding of the synthesized and real image at different layers in the network and an identity preserving network that maintains the identity of the synthesised image throughout the training process. Our approach is aimed at improving the visual appeal of the synthesised images while incorporating multiple attribute assignment to the generator without compromising the identity of the synthesised image. We synthesised sketches using XDOG filter for the CelebA, WVU Multi-modal and CelebA-HQ datasets and from an auxiliary generator trained on sketches from CUHK, IIT-D and FERET datasets. Our results are impressive compared to current state of the art.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源