Image Harmonization Guided by Semantic Information

被引:0
作者
Yang Z.-Y. [1 ,2 ]
Li P.-C. [1 ,2 ]
Liu F.-C. [1 ,2 ]
Gao C.-Q. [1 ,2 ]
机构
[1] School of Communication and Information Engineering, Chongqing University of Posts and Telecommunications, Chongqing
[2] Chongqing Key Laboratory of Signal and Information Processing, Chongqing
来源
Tien Tzu Hsueh Pao/Acta Electronica Sinica | 2023年 / 51卷 / 07期
基金
中国国家自然科学基金;
关键词
image harmonization; image processing; local background information; multi-resolution selective fusion; semantic information; spatial feature information;
D O I
10.12263/DZXB.20221322
中图分类号
学科分类号
摘要
Image harmonization occupies an important position in image processing. It aims to adjust the foreground appearance, e.g., illumination, color, texture, etc., to be visually consistent with the background. However, existing deep learning-based methods usually use the feature distribution of the overall image background as a cue to adjust the foreground, without focusing on the critical role of semantic information for foreground alignment, resulting in local areas in the foreground to appear visually different from the background. To this end, based on the multi-resolution selective fusion module (MRSFM) and the lightweight convolutional block attention module (CBAM), this paper designs a multi-resolution selective fusion module based on dual attention mechanism (MRSF-DAM), which makes the final output feature map rich in semantic information, thus guiding the network to better understand the correlation between the foreground of an image and its surrounding scene, more enabling the network to fully obtain the various information needed to coordinate the foreground from the background, and eventually reducing the visual discrepancy between the foreground and background regions of an image. In addition, this article designs a new network architecture to selectively fuse the shallow and deep feature information. By multi-scale fusion and enhancement of the output feature maps of the first six network layers of the decoder and MRSF-DAM, the generated enhanced feature maps are fed into the final layer of the decoder, which can alleviate the problem introduced by skip connections of the unrelated features to the foreground,and besides, it reduces the loss of spatial feature information caused by multiple downsampling of the decoder, further improving the authenticity of the generated harmonized images. A large number of experiments were conducted on the widely used iHarmony4 benchmark dataset to verify the effectiveness of our method. Compared to the latest method SCS Co (Self Consistent Style Comparative learning for image harmonization), this proposed method improves the mean squared error (MSE), foreground mean squared error (fMSE) and peak signal to noise ratio (PSNR) of the entire dataset by 4.28, 61.97, and 1 dB, respectively. © 2023 Chinese Institute of Electronics. All rights reserved.
引用
收藏
页码:1826 / 1834
页数:8
相关论文
共 38 条
[11]  
AN J, HUANG S Y, SONG Y B, Et al., ArtFlow: unbiased image style transfer via reversible neural flows, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 862-871, (2021)
[12]  
WU X, HU Z, SHENG L, Et al., Styleformer: Real-time arbitrary style transfer via parametric style composition, 2021 IEEE/CVF International Conference on Computer Vision (ICCV), pp. 14618-14627, (2021)
[13]  
REINHARD E, ADHIKHMIN M, GOOCH B, Et al., Color transfer between images, IEEE Computer Graphics and Applications, 21, 5, pp. 34-41, (2001)
[14]  
XIAO X Z, MA L Z., Color transfer in correlated color space, Proceedings of the 2006 ACM International Conference on Virtual Reality Continuum and Its Applications, pp. 305-309, (2006)
[15]  
FECKER U, BARKOWSKY M, KAUP A., Histogram-based prefiltering for luminance and chrominance compensation of multiview video, IEEE Transactions on Circuits and Systems for Video Technology, 18, 9, pp. 1258-1267, (2008)
[16]  
PITIE F, KOKARAM A C, DAHYOT R., Automated colour grading using colour distribution transfer, Computer Vision and Image Understanding, 107, 1, pp. 123-137, (2007)
[17]  
SUNKAVALLI K, JOHNSON M K, MATUSIK W, Et al., Multi-scale image harmonization, ACM Transactions on Graphics, 29, 4, pp. 1-10, (2010)
[18]  
SONG S B, ZHONG F, QIN X Y, Et al., Illumination Harmonization with Gray Mean Scale, Advances in Computer Graphics, pp. 193-205, (2020)
[19]  
LALONDE J F, EFROS A., Using color compatibility for assessing image realism, 2007 IEEE 11th International Conference on Computer Vision (CVPR), pp. 1-8, (2007)
[20]  
XUE S, AGARWALA A, DORSEY J, Et al., Understanding and improving the realism of image composites, ACM Transactions on Graphics, 31, 4, pp. 1-10, (2012)