Dual-Affinity Style Embedding Network for Semantic-Aligned Image Style Transfer

被引:14
|
作者
Ma, Zhuoqi [1 ]
Lin, Tianwei [2 ]
Li, Xin [2 ]
Li, Fu [2 ]
He, Dongliang [2 ]
Ding, Errui [2 ]
Wang, Nannan [3 ]
Gao, Xinbo [4 ,5 ]
机构
[1] Xidian Univ, Sch Comp Sci & Technol, Xian Key Lab Big Data & Intelligent Vis, Xian 710071, Peoples R China
[2] Baidu Inc, Dept Comp Vis Vis Technol, Beijing 100080, Peoples R China
[3] Xidian Univ, Sch Telecommun Engn, State Key Lab Integrated Serv Networks, Xian 710071, Peoples R China
[4] Xidian Univ, Sch Elect Engn, Xian 710071, Peoples R China
[5] Chongqing Univ Posts & Telecommun, Chongqing Key Lab Image Cognit, Chongqing 400065, Peoples R China
基金
中国国家自然科学基金;
关键词
Semantics; Optimization; Feature extraction; Visualization; Correlation; Training; Real-time systems; Dual-affinity; semantic style transfer; style embedding;
D O I
10.1109/TNNLS.2022.3143356
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Image style transfer aims at synthesizing an image with the content from one image and the style from another. User studies have revealed that the semantic correspondence between style and content greatly affects subjective perception of style transfer results. While current studies have made great progress in improving the visual quality of stylized images, most methods directly transfer global style statistics without considering semantic alignment. Current semantic style transfer approaches still work in an iterative optimization fashion, which is impractically computationally expensive. Addressing these issues, we introduce a novel dual-affinity style embedding network (DaseNet) to synthesize images with style aligned at semantic region granularity. In the dual-affinity module, feature correlation and semantic correspondence between content and style images are modeled jointly for embedding local style patterns according to semantic distribution. Furthermore, the semantic-weighted style loss and the region-consistency loss are introduced to ensure semantic alignment and content preservation. With the end-to-end network architecture, DaseNet can well balance visual quality and inference efficiency for semantic style transfer. Experimental results on different scene categories have demonstrated the effectiveness of the proposed method.
引用
收藏
页码:7404 / 7417
页数:14
相关论文
共 45 条
  • [41] Dual-Path Feature Aware Network for Remote Sensing Image Semantic Segmentation
    Geng, Jie
    Song, Shuai
    Jiang, Wen
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (05) : 3674 - 3686
  • [42] QRS-Trs: Style Transfer-Based Image-to-Image Translation for Carbon Stock Estimation in Quantitative Remote Sensing
    Yu, Zhenyu
    Wang, Jinnian
    Chen, Hanqing
    Yamani Idna Idris, Mohd
    IEEE ACCESS, 2025, 13 : 52726 - 52737
  • [43] Progressive Data Augmentation Method for Remote Sensing Ship Image Classification Based on Imaging Simulation System and Neural Style Transfer
    Xiao, Qi
    Liu, Bo
    Li, Zengyi
    Ni, Wei
    Yang, Zhen
    Li, Ligang
    IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2021, 14 : 9176 - 9186
  • [44] MrFDDGAN: Multireceptive Field Feature Transfer and Dual Discriminator-Driven Generative Adversarial Network for Infrared and Color Visible Image Fusion
    Li, Junwu
    Li, Binhua
    Jiang, Yaoxi
    Tian, Longwei
    Cai, Weiwei
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2023, 72
  • [45] FB-Net: Dual-Branch Foreground-Background Fusion Network With Multi-Scale Semantic Scanning for Image-Text Retrieval
    Xu, Junhao
    Liu, Zheng
    Pei, Xinlei
    Wang, Shuhuai
    Gao, Shanshan
    IEEE ACCESS, 2023, 11 : 36516 - 36537