Video Inverse Tone Mapping Network with Luma and Chroma Mapping

被引:2
作者
Huang, Peihuan [1 ]
Cao, Gaofeng [2 ,3 ]
Zhou, Fei [1 ,3 ,4 ]
Qiu, Guoping [1 ,5 ,6 ]
机构
[1] Shenzhen Univ, Coll Elect & Informat Engn, Shenzhen, Peoples R China
[2] Peking Univ, Sch Elect & Comp Engn, Beijing, Peoples R China
[3] Peng Cheng Lab, Shenzhen, Peoples R China
[4] Guangdong Key Lab Intelligent Informat Proc, Shenzhen, Peoples R China
[5] Univ Nottingham, Sch Comp Sci, Nottingham, England
[6] Guangdong Hong Kong Joint Lab Big Data Imaging &, Hong Kong, Peoples R China
来源
PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023 | 2023年
基金
中国国家自然科学基金;
关键词
Inverse Tone Mapping; Perceptual Uniformity; ICTCP color space;
D O I
10.1145/3581783.3612199
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
With the popularity of consumer high dynamic range (HDR) display devices, video inverse tone mapping (iTM) has become a research hotspot. However, existing methods are designed based on a perceptual non-uniformity color space (e.g., RGB and TCBCR), resulting in limited quality of HDR video rendered by these methods. Considering the two key factors involved in the video iTM task: luma and chroma, in this paper, we design an ICTCP color space based video iTM model, which reproduces high quality HDR video by processing luma and chroma information. Benefitting from the decorrelated perception of luma and chroma in the ICTCP color space, two global mapping networks (INet and TPNet) are developed to enhance the luma and chroma pixels, respectively. However, luma and chroma mapping in the iTM task may be affected by color appearance phenomena. Thus, a luma-chroma adaptation transform network (LCATNet) is proposed to process the luma and chroma pixels affected by color appearance phenomena, which can complement the local details to the globally enhanced luma and chroma pixels. In the LCATNet, either the luma mapping or the chroma mapping is adaptively adjusted according to both the luma and the chroma information. Besides, benefitting from the perceptually consistent property of the ICTCP color space, the same pixel errors can draw equal model attentions during the training. Thus, the proposed model can correctly render luma and chroma information without highlighting special regions or designing special training losses. Extensive experimental results demonstrate the effectiveness of the proposed model.
引用
收藏
页码:1383 / 1391
页数:9
相关论文
共 39 条
  • [11] iCAM framework for image appearance, differences, and quality
    Fairchild, MD
    Johnson, GM
    [J]. JOURNAL OF ELECTRONIC IMAGING, 2004, 13 (01) : 126 - 138
  • [12] Glorot X., 2010, P 13 INT C ART INT S, P249, DOI DOI 10.1109/LGRS.2016.2565705
  • [13] Learning a Practical SDR-to-HDRTV Up-conversion using New Dataset and Degradation Models
    Guo, Cheng
    Fan, Leidong
    Xue, Ziyu
    Jiang, Xiuhua
    [J]. 2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 22231 - 22241
  • [14] SDRTV-to-HDRTV via Hierarchical Dynamic Context Feature Mapping
    He, Gang
    Xu, Kepeng
    Xu, Li
    Wu, Chang
    Sun, Ming
    Wen, Xing
    Tai, Yu-Wing
    [J]. PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022, : 2890 - 2898
  • [15] Hunt Robert William Gainer, 2005, WILEY IS T SERIES IM
  • [16] ITU-R, 2015, Tech. Rep., ITU-R Rec, BT.2020-2, P2
  • [17] ITU-R, 2021, METH CONV HIGH DYN R
  • [18] ITU-R, 2015, COL CONV REC ITU R B
  • [19] ITU-R, 2020, HIGH DYN RANG TEL PR
  • [20] ITU-R, 2015, Rec. ITU-R BT.709-6: Parameter Values for the HDTV Standards for Production and International Programme Exchange