Rethinking Image Inpainting via a Mutual Encoder-Decoder with Feature Equalizations

被引:215
作者
Liu, Hongyu [1 ]
Jiang, Bin [1 ]
Song, Yibing [2 ]
Huang, Wei [1 ]
Yang, Chao [1 ]
机构
[1] Hunan Univ, Coll Comp Sci & Elect Engn, Changsha, Peoples R China
[2] Tencent AI Lab, Shenzhen, Peoples R China
来源
COMPUTER VISION - ECCV 2020, PT II | 2020年 / 12347卷
基金
中国国家自然科学基金;
关键词
Deep image inpainting; Feature equalizations;
D O I
10.1007/978-3-030-58536-5_43
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep encoder-decoder based CNNs have advanced image inpainting methods for hole filling. While existing methods recover structures and textures step-by-step in the hole regions, they typically use two encoder-decoders for separate recovery. The CNN features of each encoder are learned to capture either missing structures or textures without considering them as a whole. The insufficient utilization of these encoder features hampers the performance of recovering both structures and textures. In this paper, we propose a mutual encoder-decoder CNN for joint recovery of both. We use CNN features from the deep and shallow layers of the encoder to represent structures and textures of an input image, respectively. The deep layer features are sent to a structure branch, while the shallow layer features are sent to a texture branch. In each branch, we fill holes in multiple scales of the CNN features. The filled CNN features from both branches are concatenated and then equalized. During feature equalization, we reweigh channel attentions first and propose a bilateral propagation activation function to enable spatial equalization. To this end, the filled CNN features of structure and texture mutually benefit each other to represent image content at all feature levels. We then use the equalized feature to supplement decoder features for output image generation through skip connections. Experiments on benchmark datasets show that the proposed method is effective to recover structures and textures and performs favorably against state-of-the-art approaches.
引用
收藏
页码:725 / 741
页数:17
相关论文
共 43 条
[11]   Deep Residual Learning for Image Recognition [J].
He, Kaiming ;
Zhang, Xiangyu ;
Ren, Shaoqing ;
Sun, Jian .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :770-778
[12]  
Heusel M, 2017, ADV NEUR IN, V30
[13]  
Hu J, 2018, PROC CVPR IEEE, P7132, DOI [10.1109/TPAMI.2019.2913372, 10.1109/CVPR.2018.00745]
[14]   Globally and Locally Consistent Image Completion [J].
Iizuka, Satoshi ;
Simo-Serra, Edgar ;
Ishikawa, Hiroshi .
ACM TRANSACTIONS ON GRAPHICS, 2017, 36 (04)
[15]   Image-to-Image Translation with Conditional Adversarial Networks [J].
Isola, Phillip ;
Zhu, Jun-Yan ;
Zhou, Tinghui ;
Efros, Alexei A. .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :5967-5976
[16]   Perceptual Losses for Real-Time Style Transfer and Super-Resolution [J].
Johnson, Justin ;
Alahi, Alexandre ;
Li Fei-Fei .
COMPUTER VISION - ECCV 2016, PT II, 2016, 9906 :694-711
[17]  
Jolicoeur-Martineau A., 2018, ICLR
[18]  
King DB, 2015, ACS SYM SER, V1214, P1, DOI 10.1021/bk-2015-1214.ch001
[19]  
Levin A, 2003, NINTH IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION, VOLS I AND II, PROCEEDINGS, P305
[20]  
Li Y., 2017, CVPR, DOI DOI 10.1109/CVPR.2017.472