Gated-GAN: Adversarial Gated Networks for Multi-Collection Style Transfer

被引:103
作者
Chen, Xinyuan [1 ,2 ,3 ]
Xu, Chang [4 ]
Yang, Xiaokang [1 ,2 ]
Song, Li [2 ]
Tao, Dacheng [4 ]
机构
[1] Shanghai Jiao Tong Univ, MoE Key Lab Artificial Intelligence, AI Inst, Shanghai 200240, Peoples R China
[2] Shanghai Jiao Tong Univ, Inst Image Commun & Network Engn, Shanghai 200240, Peoples R China
[3] Univ Technol Sydney, Fac Engn & Informat Technol, Ctr Artificial Intelligence, Ultimo, NSW 2007, Australia
[4] Univ Sydney, Sch Informat Technol, Fac Engn & Informat Technol, UBTECH Sydney Artificial Intelligence Ctr, Darlington, NSW 2008, Australia
基金
澳大利亚研究理事会;
关键词
Multi-style transfer; adversarial generative networks; TEXTURE SYNTHESIS;
D O I
10.1109/TIP.2018.2869695
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Style transfer describes the rendering of an image's semantic content as different artistic styles. Recently, generative adversarial networks (GANs) have emerged as an effective approach in style transfer by adversarially training the generator to synthesize convincing counterfeits. However, traditional GAN suffers from the mode collapse issue, resulting in unstable training and making style transfer quality difficult to guarantee. In addition, the GAN generator is only compatible with one style, so a series of GANs must be trained to provide users with choices to transfer more than one kind of style. In this paper, we focus on tackling these challenges and limitations to improve style transfer. We propose adversarial gated networks (Gated-GAN) to transfer multiple styles in a single model. The generative networks have three modules: an encoder, a gated transformer, and a decoder. Different styles can be achieved by passing input images through different branches of the gated transformer. To stabilize training, the encoder and decoder are combined as an auto-encoder to reconstruct the input images. The discriminative networks are used to distinguish whether the input image is a stylized or genuine image. An auxiliary classifier is used to recognize the style categories of transferred images, thereby helping the generative networks generate images in multiple styles. In addition, Gated-GAN makes it possible to explore a new style by investigating styles learned from artists or genres. Our extensive experiments demonstrate the stability and effectiveness of the proposed model for multi-style transfer.
引用
收藏
页码:546 / 560
页数:15
相关论文
共 53 条
  • [1] Texture Synthesis Using the Structure Tensor
    Akl, Adib
    Yaacoub, Charles
    Donias, Marc
    Da Costa, Jean-Pierre
    Germain, Christian
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2015, 24 (11) : 4082 - 4095
  • [2] Image up-sampling using total-variation regularization with a new observation model
    Aly, HA
    Dubois, E
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2005, 14 (10) : 1647 - 1659
  • [3] [Anonymous], 2017, PROC IEEE C COMPUT V
  • [4] [Anonymous], 2017, DUALGAN UNSUPERVISED
  • [5] [Anonymous], 2017, P IEEE C COMP VIS PA
  • [6] [Anonymous], 2016, P IEEE C COMPUTER VI
  • [7] [Anonymous], PROC CVPR IEEE
  • [8] [Anonymous], TEXTURE SYNTHE SIS S
  • [9] [Anonymous], 2017, COMMUN ACM, DOI DOI 10.1145/3065386
  • [10] [Anonymous], 2017, LEARNING DISCOVER CR