论文标题

诱导条件gan的最佳属性表示

Inducing Optimal Attribute Representations for Conditional GANs

论文作者

Bhattarai, Binod, Kim, Tae-Kyun

论文摘要

有条件的gan被广泛用于将图像从一个类别转换为另一个类别。有意义的gan条件提供了更大的灵活性和控制目标域合成数据的性质。现有的条件剂量通常以0s和1s的形式编码目标域标签信息作为硬编码的分类向量。此类表示的主要缺点是无法编码目标类别的高阶语义信息及其相对依赖性。我们建议使用图形卷积网络的新颖端到端学习框架,以学习在发电机上条件的属性表示。 GAN损失,即鉴别因子和属性分类损失,被馈回图形,从而产生的合成图像在属性上更自然和更清晰。此外,在发电机侧而不是在gan的歧视侧,先前的艺术品被优先考虑条件。我们通过多任务学习也将条件应用于歧视端。我们增强了四个最先进的CGANS架构:Stargan,Stargan-Jnt,Attgan和Stgan。我们广泛的定性和定量评估对具有挑战性的面部属性操纵数据集(Celeba,LF​​WA和RAFD),表明,与他们的反零件和其他条件方法相比,我们的方法通过我们的方法优于较大的差距增强了CGAN,这两种目标属性属性识别率和质量质量率和诸如PSNR和SSSSSSSSSSSSIM和SSSIM的质量率。

Conditional GANs are widely used in translating an image from one category to another. Meaningful conditions to GANs provide greater flexibility and control over the nature of the target domain synthetic data. Existing conditional GANs commonly encode target domain label information as hard-coded categorical vectors in the form of 0s and 1s. The major drawbacks of such representations are inability to encode the high-order semantic information of target categories and their relative dependencies. We propose a novel end-to-end learning framework with Graph Convolutional Networks to learn the attribute representations to condition on the generator. The GAN losses, i.e. the discriminator and attribute classification losses, are fed back to the Graph resulting in the synthetic images that are more natural and clearer in attributes. Moreover, prior-arts are given priorities to condition on the generator side, not on the discriminator side of GANs. We apply the conditions to the discriminator side as well via multi-task learning. We enhanced the four state-of-the art cGANs architectures: Stargan, Stargan-JNT, AttGAN and STGAN. Our extensive qualitative and quantitative evaluations on challenging face attributes manipulation data set, CelebA, LFWA, and RaFD, show that the cGANs enhanced by our methods outperform by a large margin, compared to their counter-parts and other conditioning methods, in terms of both target attributes recognition rates and quality measures such as PSNR and SSIM.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源