Toward Exploiting Second-Order Feature Statistics for Arbitrary Image Style Transfer

被引:3
作者
Choi, Hyun-Chul [1 ]
机构
[1] Yeungnam Univ, Dept Elect Engn, Intelligent Comp Vis Software Lab, 280 Daehak Ro, Gyongsan 38541, Gyeongbuk, South Korea
基金
新加坡国家研究基金会;
关键词
image style transfer; second-order feature statistics; component-wise feature transform; mean and covariance loss; component-wise style control;
D O I
10.3390/s22072611
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
Generating images of artistic style from input images, also known as image style transfer, has been improved in the quality of output style and the speed of image generation since deep neural networks have been applied in the field of computer vision research. However, the previous approaches used feature alignment techniques that were too simple in their transform layer to cover the characteristics of style features of images. In addition, they used an inconsistent combination of transform layers and loss functions in the training phase to embed arbitrary styles in a decoder network. To overcome these shortcomings, the second-order statistics of the encoded features are exploited to build an optimal arbitrary image style transfer technique. First, a new correlation-aware loss and a correlation-aware feature alignment technique are proposed. Using this consistent combination of loss and feature alignment methods strongly matches the second-order statistics of content features to those of the target-style features and, accordingly, the style capacity of the decoder network is increased. Secondly, a new component-wise style controlling method is proposed. This method can generate various styles from one or several style images by using style-specific components from second-order feature statistics. We experimentally prove that the proposed method achieves improvements in both the style capacity of the decoder network and the style variety without losing the ability of real-time processing (less than 200 ms) on Graphics Processing Unit (GPU) devices.
引用
收藏
页数:12
相关论文
共 24 条
  • [1] [Anonymous], 2016, Kaggle dataset: Painter by numbers
  • [2] [Anonymous], 2017, P BRIT MACH VIS C BM
  • [3] Dumoulin V., 2017, INT C LEARN REPR
  • [4] Controlling Perceptual Factors in Neural Style Transfer
    Gatys, Leon A.
    Ecker, Alexander S.
    Bethge, Matthias
    Hertzmann, Aaron
    Shechtman, Eli
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 3730 - 3738
  • [5] Image Style Transfer Using Convolutional Neural Networks
    Gatys, Leon A.
    Ecker, Alexander S.
    Bethge, Matthias
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 2414 - 2423
  • [6] Image analogies
    Hertzmann, A
    Jacobs, CE
    Oliver, N
    Curless, B
    Salesin, DH
    [J]. SIGGRAPH 2001 CONFERENCE PROCEEDINGS, 2001, : 327 - 340
  • [7] Arbitrary Style Transfer in Real-time with Adaptive Instance Normalization
    Huang, Xun
    Belongie, Serge
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, : 1510 - 1519
  • [8] Image-to-Image Translation with Conditional Adversarial Networks
    Isola, Phillip
    Zhu, Jun-Yan
    Zhou, Tinghui
    Efros, Alexei A.
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 5967 - 5976
  • [9] Perceptual Losses for Real-Time Style Transfer and Super-Resolution
    Johnson, Justin
    Alahi, Alexandre
    Li Fei-Fei
    [J]. COMPUTER VISION - ECCV 2016, PT II, 2016, 9906 : 694 - 711
  • [10] In the light of feature distributions: moment matching for Neural Style Transfer
    Kalischek, Nikolai
    Wegner, Jan D.
    Schindler, Konrad
    [J]. 2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 9377 - 9386