Consistent Arbitrary Style Transfer Using Consistency Training and Self-Attention Module

被引:3
|
作者
Zhou, Zheng [1 ]
Wu, Yue [2 ]
Zhou, Yicong [1 ]
机构
[1] Univ Macau, Dept Comp & Informat Sci, Taipa, Macao, Peoples R China
[2] Amazon Alexa Nat Understanding, Manhattan Beach, CA 90007 USA
关键词
Image color analysis; Adaptation models; Transformers; Learning systems; Visualization; Training; Loss measurement; Arbitrary style transfer (AST); consistent training; self-attention (SA); style inconsistency;
D O I
10.1109/TNNLS.2023.3298383
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Arbitrary style transfer (AST) has garnered considerable attention for its ability to transfer styles infinitely. Although existing methods have achieved impressive results, they may overlook style consistencies and fail to capture crucial style patterns, leading to inconsistent style transfer (ST) caused by minor disturbances. To tackle this issue, we conduct a mathematical analysis of inconsistent ST and develop a style inconsistency measure (SIM) to quantify the inconsistencies between generated images. Moreover, we propose a consistent AST (CAST) framework that effectively captures and transfers essential style features into content images. The proposed CAST framework incorporates an intersection-of-union-preserving crop (IoUPC) module to obtain style pairs with minor disturbance, a self-attention (SA) module to learn the crucial style features, and a style inconsistency loss regularization (SILR) to facilitate consistent feature learning for consistent stylization. Our proposed framework not only provides an optimal solution for consistent ST but also outperforms existing methods when embedded into the CAST framework. Extensive experiments demonstrate that the proposed CAST framework can effectively transfer style patterns while preserving consistency and achieve the state-of-the-art performance.
引用
收藏
页码:16845 / 16856
页数:12
相关论文
共 50 条
  • [1] Consistent Arbitrary Style Transfer Using Consistency Training and Self-Attention Module
    Zhou, Zheng
    Wu, Yue
    Zhou, Yicong
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (11) : 16845 - 16856
  • [2] Arbitrary Style Transfer with Parallel Self-Attention
    Zhang, Tiange
    Gao, Ying
    Gao, Feng
    Qi, Lin
    Dong, Junyu
    2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 1406 - 1413
  • [3] Adversarial Training Inspired Self-attention Flow for Universal Image Style Transfer
    Dang, Kaiheng
    Lai, Jianhuang
    Dong, Junhao
    Xie, Xiaohua
    PATTERN RECOGNITION, ACPR 2021, PT II, 2022, 13189 : 476 - 489
  • [4] Arbitrary style transfer via content consistency and style consistency
    Xiaoming Yu
    Gan Zhou
    The Visual Computer, 2024, 40 : 1369 - 1382
  • [5] Arbitrary style transfer via content consistency and style consistency
    Yu, Xiaoming
    Zhou, Gan
    VISUAL COMPUTER, 2024, 40 (03): : 1369 - 1382
  • [6] FST-OAM: a fast style transfer model using optimized self-attention mechanism
    Du, Xiaozhi
    Jia, Ning
    Du, Hongyuan
    SIGNAL IMAGE AND VIDEO PROCESSING, 2024, 18 (05) : 4191 - 4203
  • [7] Semi-Supervised Sound Event Detection Using Self-Attention and Multiple Techniques of Consistency Training
    Wang, Yih-Wen
    Chen, Chia-Ping
    Lu, Chung-Li
    Chan, Bo-Cheng
    2021 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA ASC), 2021, : 269 - 274
  • [8] ESA-CycleGAN: Edge feature and self-attention based cycle-consistent generative adversarial network for style transfer
    Wang, Li
    Wang, Lidan
    Chen, Shubai
    IET IMAGE PROCESSING, 2022, 16 (01) : 176 - 190
  • [9] Indoor Depth Completion with Boundary Consistency and Self-Attention
    Huang, Yu-Kai
    Wu, Tsung-Han
    Liu, Yueh-Cheng
    Hsu, Winston H.
    2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW), 2019, : 1070 - 1078
  • [10] Adversarial training for fast arbitrary style transfer
    Xu, Zheng
    Wilber, Michael
    Fang, Chen
    Hertzmann, Aaron
    Jin, Hailin
    COMPUTERS & GRAPHICS-UK, 2020, 87 : 1 - 11