Dual-Affinity Style Embedding Network for Semantic-Aligned Image Style Transfer

被引:14
|
作者
Ma, Zhuoqi [1 ]
Lin, Tianwei [2 ]
Li, Xin [2 ]
Li, Fu [2 ]
He, Dongliang [2 ]
Ding, Errui [2 ]
Wang, Nannan [3 ]
Gao, Xinbo [4 ,5 ]
机构
[1] Xidian Univ, Sch Comp Sci & Technol, Xian Key Lab Big Data & Intelligent Vis, Xian 710071, Peoples R China
[2] Baidu Inc, Dept Comp Vis Vis Technol, Beijing 100080, Peoples R China
[3] Xidian Univ, Sch Telecommun Engn, State Key Lab Integrated Serv Networks, Xian 710071, Peoples R China
[4] Xidian Univ, Sch Elect Engn, Xian 710071, Peoples R China
[5] Chongqing Univ Posts & Telecommun, Chongqing Key Lab Image Cognit, Chongqing 400065, Peoples R China
基金
中国国家自然科学基金;
关键词
Semantics; Optimization; Feature extraction; Visualization; Correlation; Training; Real-time systems; Dual-affinity; semantic style transfer; style embedding;
D O I
10.1109/TNNLS.2022.3143356
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Image style transfer aims at synthesizing an image with the content from one image and the style from another. User studies have revealed that the semantic correspondence between style and content greatly affects subjective perception of style transfer results. While current studies have made great progress in improving the visual quality of stylized images, most methods directly transfer global style statistics without considering semantic alignment. Current semantic style transfer approaches still work in an iterative optimization fashion, which is impractically computationally expensive. Addressing these issues, we introduce a novel dual-affinity style embedding network (DaseNet) to synthesize images with style aligned at semantic region granularity. In the dual-affinity module, feature correlation and semantic correspondence between content and style images are modeled jointly for embedding local style patterns according to semantic distribution. Furthermore, the semantic-weighted style loss and the region-consistency loss are introduced to ensure semantic alignment and content preservation. With the end-to-end network architecture, DaseNet can well balance visual quality and inference efficiency for semantic style transfer. Experimental results on different scene categories have demonstrated the effectiveness of the proposed method.
引用
收藏
页码:7404 / 7417
页数:14
相关论文
共 45 条
  • [31] GAN-Based Day-to-Night Image Style Transfer for Nighttime Vehicle Detection
    Lin, Che-Tsung
    Huang, Sheng-Wei
    Wu, Yen-Yi
    Lai, Shang-Hong
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2021, 22 (02) : 951 - 963
  • [32] Text-Guided Style Transfer-Based Image Manipulation Using Multimodal Generative Models
    Togo, Ren
    Kotera, Megumi
    Ogawa, Takahiro
    Haseyama, Miki
    IEEE ACCESS, 2021, 9 : 64860 - 64870
  • [33] Speckle Noise Reduction for OCT Images Based on Image Style Transfer and Conditional GAN
    Zhou, Yi
    Yu, Kai
    Wang, Meng
    Ma, Yuhui
    Peng, Yuanyuan
    Chen, Zhongyue
    Zhu, Weifang
    Shi, Fei
    Chen, Xinjian
    IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2022, 26 (01) : 139 - 150
  • [34] Improved reversible data hiding scheme employing dual image-based least significant bit matching for secure image communication using style transfer
    Prakash, S. Jaya
    Mahalakshmi, K.
    VISUAL COMPUTER, 2022, 38 (12) : 4129 - 4150
  • [35] Image Colorization Using the Global Scene-Context Style and Pixel-Wise Semantic Segmentation
    Tram-Tran Nguyen-Quynh
    Kim, Soo-Hyung
    Nhu-Tai Do
    IEEE ACCESS, 2020, 8 : 214098 - 214114
  • [36] IFFMStyle: High-Quality Image Style Transfer Using Invalid Feature Filter Modules
    Xu, Zhijie
    Hou, Liyan
    Zhang, Jianqin
    SENSORS, 2022, 22 (16)
  • [37] MIST-Tacotron: End-to-End Emotional Speech Synthesis Using Mel-Spectrogram Image Style Transfer
    Moon, Sungwoo
    Kim, Sunghyun
    Choi, Yong-Hoon
    IEEE ACCESS, 2022, 10 : 25455 - 25463
  • [38] Semantic-Supervised Infrared and Visible Image Fusion Via a Dual-Discriminator Generative Adversarial Network
    Zhou, Huabing
    Wu, Wei
    Zhang, Yanduo
    Ma, Jiayi
    Ling, Haibin
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 635 - 648
  • [39] SSMU-Net: A Style Separation and Mode Unification Network for Multimodal Remote Sensing Image Classification
    Han, Yi
    Zhu, Hao
    Jiao, Licheng
    Yi, Xiaoyu
    Li, Xiaotong
    Hou, Biao
    Ma, Wenping
    Wang, Shuang
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2023, 61
  • [40] DDRNet: Dual-Domain Refinement Network for Remote Sensing Image Semantic Segmentation
    Yang, Zhenhao
    Bi, Fukun
    Hou, Xinghai
    Zhou, Dehao
    Wang, Yanping
    IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2024, 17 : 20177 - 20189