Structure-Guided Arbitrary Style Transfer for Artistic Image and Video

被引:30
作者
Liu, Shiguang [1 ,2 ]
Zhu, Ting [1 ]
机构
[1] Tianjin Univ, Coll Intelligence & Comp, Tianjin 300350, Peoples R China
[2] Tianjin Univ, Tianjin Key Lab Cognit Comp & Applicat, Tianjin 300350, Peoples R China
关键词
Image edge detection; Feature extraction; Distortion; Visualization; Optimization; Graphical models; Distribution functions; Image and video style transfer; refine network; region structure loss; cycle-temporal loss; structure-guided;
D O I
10.1109/TMM.2021.3063605
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Recently, neural style transfer has become a popular task in both academic research and industrial applications. Although the existing methods made great progress in terms of quality and efficiency, most of them mainly focus on extracting high-level features. Therefore, it is still challenging to display the hierarchical structure of the content image due to lack of texture information, which causes blurred boundaries and distortion of the stylized image. In this paper, a novel neural image and video style transfer scheme is proposed to suppress distortion and preserve the semantic content of the content image, which is capable of yielding satisfactory stylized images and videos of a variety of scenarios. We first propose to assemble a refine network into an auto-encoder framework to guide style transfer, which can ensure that the stylized image have diverse levels of details. Then, we introduce the global content loss and the local region structure loss to train the model and enhance the robustness of the model. In addition, in order to produce a high-quality stylized video, our method not only preserves the image structure, but also introduces a temporal consistency loss and a cycle-temporal loss to avoid temporal incoherence and motion blur as far as possible. Our approach is also friendly for photographic and exposed image and video style transfer. Both quantitative and qualitative evaluation demonstrated the effectiveness of our method.
引用
收藏
页码:1299 / 1312
页数:14
相关论文
共 50 条
[1]   Unpaired Motion Style Transfer from Video to Animation [J].
Aberman, Kfir ;
Weng, Yijia ;
Lischinski, Dani ;
Cohen-Or, Daniel ;
Chen, Baoquan .
ACM TRANSACTIONS ON GRAPHICS, 2020, 39 (04)
[2]  
[Anonymous], 2014, COMPUT RES REPOSITOR
[3]  
[Anonymous], 2018, INT J COMPUT VISION, DOI DOI 10.1007/s11263-018-1089-z
[4]   High-for-Low and Low-for-High: Efficient Boundary Detection from Deep Object Features and its Applications to High-Level Vision [J].
Bertasius, Gedas ;
Shi, Jianbo ;
Torresani, Lorenzo .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, :504-512
[5]  
Bertasius G, 2015, PROC CVPR IEEE, P4380, DOI 10.1109/CVPR.2015.7299067
[6]   Coherent Online Video Style Transfer [J].
Chen, Dongdong ;
Liao, Jing ;
Yuan, Lu ;
Yu, Nenghai ;
Hua, Gang .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, :1114-1123
[7]   StyleBank: An Explicit Representation for Neural Image Style Transfer [J].
Chen, Dongdong ;
Yuan, Lu ;
Liao, Jing ;
Yu, Nenghai ;
Hua, Gang .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :2770-2779
[8]   Quality-Aware Unpaired Image-to-Image Translation [J].
Chen, Lei ;
Wu, Le ;
Hu, Zhenzhen ;
Wang, Meng .
IEEE TRANSACTIONS ON MULTIMEDIA, 2019, 21 (10) :2664-2674
[9]   Living arrangements and intergenerational monetary transfers of older Chinese [J].
Chen, Taichang ;
Leeson, George W. ;
Liu, Changping .
AGEING & SOCIETY, 2017, 37 (09) :1798-1823
[10]  
Chen W, 2016, ADV NEUR IN, V29