SDP-GAN: Saliency Detail Preservation Generative Adversarial Networks for High Perceptual Quality Style Transfer

被引:25
作者
Li, Ru [1 ]
Wu, Chi-Hao [2 ]
Liu, Shuaicheng [1 ]
Wang, Jue [2 ]
Wang, Guangfu [2 ]
Liu, Guanghui [1 ]
Zeng, Bing [1 ]
机构
[1] Univ Elect Sci & Technol China, Sch Informat & Commun Engn, Chengdu 611731, Peoples R China
[2] Megvii Technol, Chengdu 610000, Peoples R China
基金
中国国家自然科学基金;
关键词
Generative adversarial network; style transfer; detail preservation;
D O I
10.1109/TIP.2020.3036754
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The paper proposes a solution to effectively handle salient regions for style transfer between unpaired datasets. Recently, Generative Adversarial Networks (GAN) have demonstrated their potentials of translating images from source domain X to target domain Y in the absence of paired examples. However, such a translation cannot guarantee to generate high perceptual quality results. Existing style transfer methods work well with relatively uniform content, they often fail to capture geometric or structural patterns that always belong to salient regions. Detail losses in structured regions and undesired artifacts in smooth regions are unavoidable even if each individual region is correctly transferred into the target style. In this paper, we propose SDP-GAN, a GAN-based network for solving such problems while generating enjoyable style transfer results. We introduce a saliency network, which is trained with the generator simultaneously. The saliency network has two functions: (1) providing constraints for content loss to increase punishment for salient regions, and (2) supplying saliency features to generator to produce coherent results. Moreover, two novel losses are proposed to optimize the generator and saliency networks. The proposed method preserves the details on important salient regions and improves the total image perceptual quality. Qualitative and quantitative comparisons against several leading prior methods demonstrates the superiority of our method.
引用
收藏
页码:374 / 385
页数:12
相关论文
共 56 条
  • [11] Structure-Preserving Neural Style Transfer
    Cheng, Ming-Ming
    Liu, Xiao-Chang
    Wang, Jie
    Lu, Shao-Ping
    Lai, Yu-Kun
    Rosin, Paul L.
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 : 909 - 920
  • [12] StarGAN: Unified Generative Adversarial Networks for Multi-Domain Image-to-Image Translation
    Choi, Yunjey
    Choi, Minje
    Kim, Munyoung
    Ha, Jung-Woo
    Kim, Sunghun
    Choo, Jaegul
    [J]. 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 8789 - 8797
  • [13] Dumoulin V., 2016, ICLR
  • [14] Predicting Depth, Surface Normals and Semantic Labels with a Common Multi-Scale Convolutional Architecture
    Eigen, David
    Fergus, Rob
    [J]. 2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, : 2650 - 2658
  • [15] Task Specific Visual Saliency Prediction with Memory Augmented Conditional Generative Adversarial Networks
    Fernando, Tharindu
    Denman, Simon
    Sridharan, Sridha
    Fookes, Clinton
    [J]. 2018 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2018), 2018, : 1539 - 1548
  • [16] Gatys L., 2016, arXiv, V16, P326, DOI [DOI 10.1167/16.12.326, 10.1167/16.12.326]
  • [17] Controlling Perceptual Factors in Neural Style Transfer
    Gatys, Leon A.
    Ecker, Alexander S.
    Bethge, Matthias
    Hertzmann, Aaron
    Shechtman, Eli
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 3730 - 3738
  • [18] Image Style Transfer Using Convolutional Neural Networks
    Gatys, Leon A.
    Ecker, Alexander S.
    Bethge, Matthias
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 2414 - 2423
  • [19] Deep Residual Learning for Image Recognition
    He, Kaiming
    Zhang, Xiangyu
    Ren, Shaoqing
    Sun, Jian
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 770 - 778
  • [20] Hensel M, 2017, ADV NEUR IN, V30