Controlling strokes in fast neural style transfer using content transforms

被引:6
作者
Reimann, Max [1 ]
Buchheim, Benito [1 ]
Semmo, Amir [2 ]
Doellner, Juergen [1 ]
Trapp, Matthias [1 ]
机构
[1] Univ Potsdam, Hasso Plattner Inst, Potsdam, Germany
[2] DigitalMasterpieces GmbH, R&D, Potsdam, Germany
关键词
Open Access funding enabled and organized by Projekt DEAL. This work was partially funded by the German Federal Ministry of Education and Research (BMBF) through grants 01IS18092 (mdViPro) and 01IS19006 (KI-LAB-ITSE);
D O I
10.1007/s00371-022-02518-x
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Fast style transfer methods have recently gained popularity in art-related applications as they make a generalized real-time stylization of images practicable. However, they are mostly limited to one-shot stylizations concerning the interactive adjustment of style elements. In particular, the expressive control over stroke sizes or stroke orientations remains an open challenge. To this end, we propose a novel stroke-adjustable fast style transfer network that enables simultaneous control over the stroke size and intensity, and allows a wider range of expressive editing than current approaches by utilizing the scale-variance of convolutional neural networks. Furthermore, we introduce a network-agnostic approach for style-element editing by applying reversible input transformations that can adjust strokes in the stylized output. At this, stroke orientations can be adjusted, and warping-based effects can be applied to stylistic elements, such as swirls or waves. To demonstrate the real-world applicability of our approach, we present StyleTune, a mobile app for interactive editing of neural style transfers at multiple levels of control. Our app allows stroke adjustments on a global and local level. It furthermore implements an on-device patch-based upsampling step that enables users to achieve results with high output fidelity and resolutions of more than 20 megapixels. Our approach allows users to art-direct their creations and achieve results that are not possible with current style transfer applications.
引用
收藏
页码:4019 / 4033
页数:15
相关论文
共 49 条
[1]  
Amato G, 2019, Arxiv, DOI [arXiv:1905.04175, 10.48550/arXiv.1905.04175, DOI 10.48550/ARXIV.1905.04175]
[2]  
[Anonymous], 2015, P 3 INT C LEARN REPR
[3]  
[Anonymous], 2016, P INT S NONPH AN REN, DOI DOI 10.2312/EXP.20161067
[4]  
Babaeizadeh M., 2020, 8 INT C LEARNING REP
[5]   PatchTable: Efficient Patch Queries for Large Datasets and Applications [J].
Barnes, Connelly ;
Zhang, Fang-Lue ;
Lou, Liming ;
Wu, Xian ;
Hu, Shi-Min .
ACM TRANSACTIONS ON GRAPHICS, 2015, 34 (04)
[6]   PatchMatch: A Randomized Correspondence Algorithm for Structural Image Editing [J].
Barnes, Connelly ;
Shechtman, Eli ;
Finkelstein, Adam ;
Goldman, Dan B. .
ACM TRANSACTIONS ON GRAPHICS, 2009, 28 (03)
[8]   Stereoscopic Neural Style Transfer [J].
Chen, Dongdong ;
Yuan, Lu ;
Liao, Jing ;
Yu, Nenghai ;
Hua, Gang .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :6654-6663
[9]  
Dapkus D., How to transfer styles to images with Adobe Pho- toshop
[10]  
Dumoulin V., 2017, P INT C LEARN REPR