Shift-Net: Image Inpainting via Deep Feature Rearrangement

被引:353
作者
Yan, Zhaoyi [1 ]
Li, Xiaoming [1 ]
Li, Mu [2 ]
Zuo, Wangmeng [1 ]
Shan, Shiguang [3 ]
机构
[1] Harbin Inst Technol, Sch Comp Sci & Technol, Harbin, Peoples R China
[2] Hong Kong Polytech Univ, Dept Comp, Hung Hom, Hong Kong, Peoples R China
[3] Chinese Acad Sci, Inst Comp Technol, Beijing 100049, Peoples R China
来源
COMPUTER VISION - ECCV 2018, PT XIV | 2018年 / 11218卷
基金
中国国家自然科学基金;
关键词
Inpainting; Feature rearrangement; Deep learning; COMPLETION;
D O I
10.1007/978-3-030-01264-9_1
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep convolutional networks (CNNs) have exhibited their potential in image inpainting for producing plausible results. However, in most existing methods, e.g., context encoder, the missing parts are predicted by propagating the surrounding convolutional features through a fully connected layer, which intends to produce semantically plausible but blurry result. In this paper, we introduce a special shift-connection layer to the U-Net architecture, namely Shift-Net, for filling in missing regions of any shape with sharp structures and fine-detailed textures. To this end, the encoder feature of the known region is shifted to serve as an estimation of the missing parts. A guidance loss is introduced on decoder feature to minimize the distance between the decoder feature after fully connected layer and the ground-truth encoder feature of the missing parts. With such constraint, the decoder feature in missing region can be used to guide the shift of encoder feature in known region. An end-to-end learning algorithm is further developed to train the Shift-Net. Experiments on the Paris StreetView and Places datasets demonstrate the efficiency and effectiveness of our Shift-Net in producing sharper, fine-detailed, and visually plausible results. The codes and pre-trained models are available at https://github.com/Zhaoyi-Yan/Shift-Net.
引用
收藏
页码:3 / 19
页数:17
相关论文
共 44 条
[1]  
[Anonymous], 2006, 2006 IEEE COMP SOC C
[2]  
[Anonymous], 2010, LECT NOTES COMPUT SC
[3]  
[Anonymous], ARXIV161107865
[4]  
[Anonymous], 2016, P 33 INT C INT C MAC
[5]  
[Anonymous], 2015, PROC INT C NEURAL IN
[6]   PatchMatch: A Randomized Correspondence Algorithm for Structural Image Editing [J].
Barnes, Connelly ;
Shechtman, Eli ;
Finkelstein, Adam ;
Goldman, Dan B. .
ACM TRANSACTIONS ON GRAPHICS, 2009, 28 (03)
[7]   Living arrangements and intergenerational monetary transfers of older Chinese [J].
Chen, Taichang ;
Leeson, George W. ;
Liu, Changping .
AGEING & SOCIETY, 2017, 37 (09) :1798-1823
[8]  
CRIMINISI A, 2003, PROC CVPR IEEE, P721, DOI [DOI 10.1109/CVPR.2003.1211538, 10. 1109/CVPR.2003.1211538]
[9]   What Makes Paris Look like Paris? [J].
Doersch, Carl ;
Singh, Saurabh ;
Gupta, Abhinav ;
Sivic, Josef ;
Efros, Alexei A. .
ACM TRANSACTIONS ON GRAPHICS, 2012, 31 (04)
[10]   Fragment-based image completion [J].
Drori, I ;
Cohen-Or, D ;
Yeshurun, H .
ACM TRANSACTIONS ON GRAPHICS, 2003, 22 (03) :303-312