Cascading Blend Network for Image Inpainting

被引:9
作者
Jin, Yiting [1 ]
Wu, Jie [1 ]
Wang, Wanliang [1 ]
Yan, Yidong [1 ]
Jiang, Jiawei [1 ]
Zheng, Jianwei [1 ]
机构
[1] Zhejiang Univ Technol, Coll Comp Sci & Technol, Liuhe Rd, Hangzhou 310023, Zhejiang, Peoples R China
基金
中国国家自然科学基金;
关键词
Image inpainting; deep convolutional networks; large-proportion corrupted images; attention-based; multi-scale context blend;
D O I
10.1145/3608952
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Image inpainting refers to filling in unknown regions with known knowledge, which is in full flourish accompanied by the popularity and prosperity of deep convolutional networks. Current inpainting methods have excelled in completing small-sized corruption or specifically masked images. However, for large-proportion corrupted images, most attention-based and structure-based approaches, though reported with state-of-the-art performance, fail to reconstruct high-quality results due to the short consideration of semantic relevance. To relieve the above problem, in this paper, we propose a novel image inpainting approach, namely cascading blend network (CBNet), to strengthen the capacity of feature representation. As a whole, we introduce an adjacent transfer attention (ATA) module in the decoder, which preserves contour structure reasonably from the deep layer and blends structure-texture information from the shadow layer. In a coarse to delicate manner, a multi-scale contextual blend (MCB) block is further designed to felicitously assemble the multi-stage feature information. In addition, to ensure a high qualified hybrid of the feature information, extra deep supervision is applied to the intermediate features through a cascaded loss. Qualitative and quantitative experiments on the Paris StreetView, CelebA, and Places2 datasets demonstrate the superior performance of our approach compared with most state-of-the-art algorithms.
引用
收藏
页数:21
相关论文
共 69 条
[11]   Structure-aware Meta-fusion for Image Super-resolution [J].
Ma H. ;
Gong B. ;
Yu Y. .
ACM Transactions on Multimedia Computing, Communications and Applications, 2022, 18 (02)
[12]  
Ho Jonathan., 2020, ADV NEURAL INFORM PR, V33, P6840
[13]   Sketch-guided Deep Portrait Generation [J].
Ho, Trang-Thi ;
Virtusio, John Jethro ;
Chen, Yung-Yao ;
Hsu, Chih-Ming ;
Hua, Kai-Lung .
ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2020, 16 (03)
[14]   Rethinking Image Inpainting via a Mutual Encoder-Decoder with Feature Equalizations [J].
Liu, Hongyu ;
Jiang, Bin ;
Song, Yibing ;
Huang, Wei ;
Yang, Chao .
COMPUTER VISION - ECCV 2020, PT II, 2020, 12347 :725-741
[15]   An efficient nonlocal variational method with application to underwater image restoration [J].
Hou, Guojia ;
Pan, Zhenkuan ;
Wang, Guodong ;
Yang, Huan ;
Duan, Jinming .
NEUROCOMPUTING, 2019, 369 :106-121
[16]   Globally and Locally Consistent Image Completion [J].
Iizuka, Satoshi ;
Simo-Serra, Edgar ;
Ishikawa, Hiroshi .
ACM TRANSACTIONS ON GRAPHICS, 2017, 36 (04)
[17]  
Jiang K, 2022, Arxiv, DOI arXiv:2207.10455
[18]   Multi-Scale Progressive Fusion Network for Single Image Deraining [J].
Jiang, Kui ;
Wang, Zhongyuan ;
Yi, Peng ;
Chen, Chen ;
Huang, Baojin ;
Luo, Yimin ;
Ma, Jiayi ;
Jiang, Junjun .
2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2020), 2020, :8343-8352
[19]   Dense Vehicle Counting Estimation via a Synergism Attention Network [J].
Jin, Yiting ;
Wu, Jie ;
Wang, Wanliang ;
Wang, Yibin ;
Yang, Xi ;
Zheng, Jianwei .
ELECTRONICS, 2022, 11 (22)
[20]   Chinese Character Image Completion Using a Generative Latent Variable Model [J].
Jo, In-su ;
Choi, Dong-bin ;
Park, Young B. .
APPLIED SCIENCES-BASEL, 2021, 11 (02) :1-19