Uncorrelated feature encoding for faster image style transfer

被引:16
作者
Kim, Minseong [1 ]
Choi, Hyun-Chul [2 ]
机构
[1] Alchera Inc, 225-15 Pangyoyeok Ro, Seongnam 13494, Gyeonggi Do, South Korea
[2] Yeungnam Univ, ICVSLab, Dept Elect Engn, 280 Daehak Ro, Gyongsan 38541, Gyeongbuk, South Korea
基金
新加坡国家研究基金会;
关键词
Uncorrelated feature encoding; Uncorrelation loss; Image style transfer; Convolutional neural networks; End-to-end learning; Redundant channel elimination;
D O I
10.1016/j.neunet.2021.03.007
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent image style transfer methods use a pre-trained convolutional neural network as their feature encoder. However, the pre-trained network is not optimal for image style transfer but rather for image classification. Furthermore, they require time-consuming feature alignment to consider the existing correlation among channels of the encoded feature map. In this paper, we propose an end-to-end learning method that optimizes both encoder and decoder networks for style transfer task and relieves the computational complexity of the existing correlation-aware feature alignment. First, we performed end-to-end learning that updates not only decoder but also encoder parameters for the task of image style transfer in the network training phase. Second, in addition to the previous style and content losses, we use uncorrelation loss, i.e., the total correlation coefficient among responses of encoder channels. Our uncorrelation loss allows the encoder network to generate a feature map of channels without correlation. Subsequently, our method results in faster forward processing with only a light-weighted transformer of correlation-unaware feature alignment. Moreover, our method drastically reduced the channel redundancy of the encoded feature during the network training process. This provides us a possibility to perform channel elimination with negligible degradation in generated style quality. Our method is applicable to multiple scaled style transfer by using the cascade network scheme and allows a user to control style strength through the usage of a content-style trade-off parameter. (C) 2021 Elsevier Ltd. All rights reserved.
引用
收藏
页码:148 / 157
页数:10
相关论文
共 28 条
  • [1] [Anonymous], 2016, Kaggle dataset: Painter by numbers
  • [2] [Anonymous], 2017, BRIT MACH VIS C
  • [3] [Anonymous], 2017, IEEE C COMP VIS PATT
  • [4] StyleBank: An Explicit Representation for Neural Image Style Transfer
    Chen, Dongdong
    Yuan, Lu
    Liao, Jing
    Yu, Nenghai
    Hua, Gang
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 2770 - 2779
  • [5] Cogswell M., 2016, INT C LEARN REPR
  • [6] Dumoulin V., 2017, ICLR
  • [7] Controlling Perceptual Factors in Neural Style Transfer
    Gatys, Leon A.
    Ecker, Alexander S.
    Bethge, Matthias
    Hertzmann, Aaron
    Shechtman, Eli
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 3730 - 3738
  • [8] Image Style Transfer Using Convolutional Neural Networks
    Gatys, Leon A.
    Ecker, Alexander S.
    Bethge, Matthias
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 2414 - 2423
  • [9] Goodfellow I., 2020, ADV NEUR IN, V63, P139, DOI [DOI 10.1145/3422622, 10.1145/3422622]
  • [10] Arbitrary Style Transfer in Real-time with Adaptive Instance Normalization
    Huang, Xun
    Belongie, Serge
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, : 1510 - 1519