Exploring the Temporal Consistency of Arbitrary Style Transfer: A Channelwise Perspective

被引:17
|
作者
Kong, Xiaoyu [1 ,2 ]
Deng, Yingying [3 ]
Tang, Fan [4 ]
Dong, Weiming [3 ]
Ma, Chongyang [5 ]
Chen, Yongyong
He, Zhenyu [6 ,7 ]
Xu, Changsheng [3 ]
机构
[1] Jilin Univ, Sch Artificial Intelligence, Changchun 130012, Peoples R China
[2] Harbin Inst Technol, Sch Comp Sci & Technol, Shenzhen 518073, Peoples R China
[3] Chinese Acad Sci, Inst Automat, Natl Lab Pattern Recognit, Beijing 100190, Peoples R China
[4] Chinese Acad Sci, Inst Comp Technol, Beijing 100190, Peoples R China
[5] Kuaishou Technol, Beijing 100085, Peoples R China
[6] Harbin Inst Technol, Dept Comp Sci, Shenzhen 518073, Peoples R China
[7] Peng Cheng Lab, Shenzhen 518055, Peoples R China
基金
美国国家科学基金会; 国家重点研发计划;
关键词
Correlation; Task analysis; Optical imaging; Integrated optics; Lighting; Optical fiber networks; Image reconstruction; Arbitrary stylization; channel correlation; cross-domain; feature migration;
D O I
10.1109/TNNLS.2022.3230084
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Arbitrary image stylization by neural networks has become a popular topic, and video stylization is attracting more attention as an extension of image stylization. However, when image stylization methods are applied to videos, unsatisfactory results that suffer from severe flickering effects appear. In this article, we conducted a detailed and comprehensive analysis of the cause of such flickering effects. Systematic comparisons among typical neural style transfer approaches show that the feature migration modules for state-of-the-art (SOTA) learning systems are ill-conditioned and could lead to a channelwise misalignment between the input content representations and the generated frames. Unlike traditional methods that relieve the misalignment via additional optical flow constraints or regularization modules, we focus on keeping the temporal consistency by aligning each output frame with the input frame. To this end, we propose a simple yet efficient multichannel correlation network (MCCNet), to ensure that output frames are directly aligned with inputs in the hidden feature space while maintaining the desired style patterns. An inner channel similarity loss is adopted to eliminate side effects caused by the absence of nonlinear operations such as softmax for strict alignment. Furthermore, to improve the performance of MCCNet under complex light conditions, we introduce an illumination loss during training. Qualitative and quantitative evaluations demonstrate that MCCNet performs well in arbitrary video and image style transfer tasks.
引用
收藏
页码:8482 / 8496
页数:15
相关论文
共 50 条
  • [31] Arbitrary Style Transfer with Parallel Self-Attention
    Zhang, Tiange
    Gao, Ying
    Gao, Feng
    Qi, Lin
    Dong, Junyu
    2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 1406 - 1413
  • [32] Arbitrary Motion Style Transfer via Contrastive Learning
    Yang, Zirui
    Luo, Zhongwen
    Wang, Zhengrui
    Liu, Yuanyuan
    APPLIED SCIENCES-BASEL, 2025, 15 (04):
  • [33] LCCStyle: Arbitrary Style Transfer With Low Computational Complexity
    Huang, Yujie
    Jing, Minge
    Zhou, Jinjia
    Liu, Yuhao
    Fan, Yibo
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 501 - 514
  • [34] Frequency Domain Disentanglement for Arbitrary Neural Style Transfer
    Li, Dongyang
    Luo, Hao
    Wang, Pichao
    Wang, Zhibin
    Liu, Shang
    Wang, Fan
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 1, 2023, : 1287 - 1295
  • [35] Improving the Transferability of Adversarial Examples with Arbitrary Style Transfer
    Ge, Zhijin
    Shang, Fanhua
    Liu, Hongying
    Liu, Yuanyuan
    Wan, Liang
    Feng, Wei
    Wang, Xiaosen
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 4440 - 4449
  • [36] All-to-key Attention for Arbitrary Style Transfer
    Zhu, Mingrui
    He, Xiao
    Wang, Nannan
    Wang, Xiaoyu
    Gao, Xinbo
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 23052 - 23062
  • [37] A SALIENCY-AWARE METHOD FOR ARBITRARY STYLE TRANSFER
    Li, Yitong
    Ma, Jie
    2023 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2023, : 740 - 744
  • [38] ETNet: Error Transition Network for Arbitrary Style Transfer
    Song, Chunjin
    Wu, Zhijie
    Zhou, Yang
    Gong, Minglun
    Huang, Hui
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [39] ARBITRARY STYLE TRANSFER USING GRAPH INSTANCE NORMALIZATION
    Jung, Dongki
    Yang, Seunghan
    Choi, Jaehoon
    Kim, Changick
    2020 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2020, : 1596 - 1600
  • [40] QR code arbitrary style transfer algorithm based on style matching layer
    Li, Hai-Sheng
    Chen, Jingyin
    Huang, Huafeng
    MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 83 (13) : 38505 - 38522