SDP-GAN: Saliency Detail Preservation Generative Adversarial Networks for High Perceptual Quality Style Transfer

被引:25
作者
Li, Ru [1 ]
Wu, Chi-Hao [2 ]
Liu, Shuaicheng [1 ]
Wang, Jue [2 ]
Wang, Guangfu [2 ]
Liu, Guanghui [1 ]
Zeng, Bing [1 ]
机构
[1] Univ Elect Sci & Technol China, Sch Informat & Commun Engn, Chengdu 611731, Peoples R China
[2] Megvii Technol, Chengdu 610000, Peoples R China
基金
中国国家自然科学基金;
关键词
Generative adversarial network; style transfer; detail preservation;
D O I
10.1109/TIP.2020.3036754
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The paper proposes a solution to effectively handle salient regions for style transfer between unpaired datasets. Recently, Generative Adversarial Networks (GAN) have demonstrated their potentials of translating images from source domain X to target domain Y in the absence of paired examples. However, such a translation cannot guarantee to generate high perceptual quality results. Existing style transfer methods work well with relatively uniform content, they often fail to capture geometric or structural patterns that always belong to salient regions. Detail losses in structured regions and undesired artifacts in smooth regions are unavoidable even if each individual region is correctly transferred into the target style. In this paper, we propose SDP-GAN, a GAN-based network for solving such problems while generating enjoyable style transfer results. We introduce a saliency network, which is trained with the generator simultaneously. The saliency network has two functions: (1) providing constraints for content loss to increase punishment for salient regions, and (2) supplying saliency features to generator to produce coherent results. Moreover, two novel losses are proposed to optimize the generator and saliency networks. The proposed method preserves the details on important salient regions and improves the total image perceptual quality. Qualitative and quantitative comparisons against several leading prior methods demonstrates the superiority of our method.
引用
收藏
页码:374 / 385
页数:12
相关论文
共 56 条
  • [1] [Anonymous], 2017, arXiv preprint arXiv:1701.08893
  • [2] [Anonymous], 2001, Non-Photorealistic Rendering
  • [3] [Anonymous], 2017, ARXIV170101081
  • [4] [Anonymous], 2017, CORR
  • [5] [Anonymous], 2018, ARXIV180305573
  • [6] Borji A, 2012, PROC CVPR IEEE, P438, DOI 10.1109/CVPR.2012.6247706
  • [7] StyleBank: An Explicit Representation for Neural Image Style Transfer
    Chen, Dongdong
    Yuan, Lu
    Liao, Jing
    Yu, Nenghai
    Hua, Gang
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 2770 - 2779
  • [8] Gated-GAN: Adversarial Gated Networks for Multi-Collection Style Transfer
    Chen, Xinyuan
    Xu, Chang
    Yang, Xiaokang
    Song, Li
    Tao, Dacheng
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2019, 28 (02) : 546 - 560
  • [9] Chen Y.-L., 2016, P BRIT MACH VIS C, P1
  • [10] CartoonGAN: Generative Adversarial Networks for Photo Cartoonization
    Chen, Yang
    Lai, Yu-Kun
    Liu, Yong-Jin
    [J]. 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 9465 - 9474