End-to-end learning for arbitrary image style transfer

被引:5
|
作者
Yoon, Y. B. [1 ]
Kim, M. S. [2 ]
Choi, H. C. [2 ]
机构
[1] POSTECH, Dept EE, Pohang, Gyeongbuk, South Korea
[2] Yeungnam Univ, Dept Elect Engn, Gyongsan 712749, Gyeongbuk, South Korea
基金
新加坡国家研究基金会;
关键词
encoding; image classification; optimisation; stylised images; real-time arbitrary style transfer; feed-forward network; pre-trained encoder; trainable decoder; style quality; output image; image classification task; end-to-end learning scheme; fixed encoder; arbitrary image style transfer;
D O I
10.1049/el.2018.6497
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Real-time arbitrary style transfer is based on a feed-forward network, which consists of a pre-trained encoder, a feature transformer, and a trainable decoder. However, the previous approach has some degrade in style quality of output image because the pre-trained encoder is not optimised for image style transfer but originally for image classification task. An end-to-end learning scheme is introduced that optimises the encoder as well as the decoder for the task of arbitrary image style transfer. Experiments conducted with a public database proves that the style transfer network trained with the end-to-end learning scheme outperforms the network with a fixed encoder in terms of minimising both content and style losses and quality of the stylised images.
引用
收藏
页码:1276 / 1277
页数:2
相关论文
共 50 条
  • [1] LEARNING LATENT REPRESENTATIONS FOR STYLE CONTROL AND TRANSFER IN END-TO-END SPEECH SYNTHESIS
    Zhang, Ya-Jie
    Pan, Shifeng
    He, Lei
    Ling, Zhen-Hua
    2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 6945 - 6949
  • [2] IMPROVING UNSUPERVISED STYLE TRANSFER IN END-TO-END SPEECH SYNTHESIS WITH END-TO-END SPEECH RECOGNITION
    Liu, Da-Rong
    Yang, Chi-Yu
    Wu, Szu-Lin
    Lee, Hung-Yi
    2018 IEEE WORKSHOP ON SPOKEN LANGUAGE TECHNOLOGY (SLT 2018), 2018, : 640 - 647
  • [3] Data Augmentation Technology Driven By Image Style Transfer in Self-Driving Car Based on End-to-End Learning
    Liu, Dongjie
    Zhao, Jin
    Xi, Axin
    Wang, Chao
    Huang, Xinnian
    Lai, Kuncheng
    Liu, Chang
    CMES-COMPUTER MODELING IN ENGINEERING & SCIENCES, 2020, 122 (02): : 593 - 617
  • [4] End-to-End Versatile Human Activity Recognition with Activity Image Transfer Learning
    Ye, Yalan
    Liu, Ziqi
    Huang, Ziwei
    Pan, Tongjie
    Wan, Zhengyi
    2021 43RD ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE & BIOLOGY SOCIETY (EMBC), 2021, : 1128 - 1131
  • [5] End-to-End Learning for Image Burst Deblurring
    Wieschollek, Patrick
    Schoelkopf, Bernhard
    Lensch, Hendrik P. A.
    Hirsch, Michael
    COMPUTER VISION - ACCV 2016, PT IV, 2017, 10114 : 35 - 51
  • [6] End-to-end Learning for Encrypted Image Retrieval
    Feng, Qihua
    Li, Peiya
    Lu, ZhiXun
    Liu, Guan
    Huang, Feiran
    2021 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA ASC), 2021, : 1839 - 1845
  • [7] END-TO-END LYRICS RECOGNITION WITH VOICE TO SINGING STYLE TRANSFER
    Basak, Sakya
    Agarwal, Shrutina
    Ganapathy, Sriram
    Takahashi, Naoya
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 266 - 270
  • [8] Style Tokens: Unsupervised Style Modeling, Control and Transfer in End-to-End Speech Synthesis
    Wang, Yuxuan
    Stanton, Daisy
    Zhang, Yu
    Skerry-Ryan, R. J.
    Battenberg, Eric
    Shor, Joel
    Xiao, Ying
    Ren, Fei
    Jia, Ye
    Saurous, Rif A.
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 80, 2018, 80
  • [9] Learning End-to-End Lossy Image Compression: A Benchmark
    Hu, Yueyu
    Yang, Wenhan
    Ma, Zhan
    Liu, Jiaying
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (08) : 4194 - 4211
  • [10] A FAST END-TO-END METHOD WITH STYLE TRANSFER FOR ROOM LAYOUT ESTIMATION
    Chen, Junming
    Shao, Tie
    Zhang, Dongyang
    Wu, Xuehui
    2019 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2019, : 964 - 969