A Novel Just-Noticeable-Difference-Based Saliency-Channel Attention Residual Network for Full-Reference Image Quality Predictions

被引:31
|
作者
Seo, Soomin [1 ]
Ki, Sehwan [1 ]
Kim, Munchurl [1 ]
机构
[1] Korea Adv Inst Sci & Technol, Sch Elect Engn, Daejeon 34141, South Korea
关键词
Image quality; Visualization; Sensitivity; Predictive models; Distortion; Feature extraction; Visual systems; Image quality assessment (IQA); human visual systems (HVS); just noticeable difference (JND); saliency map; convolutional neural network (CNN); spatial and channel attention; FREE-ENERGY PRINCIPLE; SIMILARITY; INDEX; INFORMATION; DEVIATION; EFFICIENT;
D O I
10.1109/TCSVT.2020.3030895
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Recently, due to the strength of deep convolutional neural networks (CNN), many CNN-based image quality assessment (IQA) models have been studied. However, previous CNN-based IQA models likely have yet to utilize the characteristics of the human visual system (HVS) fully for IQA problems when they simply entrust everything to the CNN, expecting it to learn from a training dataset. Therefore, the performance capabilities of such deep-learning-based methods are somewhat saturated. However, in this article, we propose a novel saliency-channel attention residual network based on the just-noticeable-difference (JND) concept for full-reference image quality assessments (FR-IQA). It is referred to as JND-SalCAR and shows significant improvements in large IQA datasets with various types of distortion. The proposed JND-SalCAR effectively learns how to incorporate human psychophysical characteristics, such as visual saliency and JND, into image quality predictions. In the proposed network, a SalCAR block is devised so that perceptually important features can be extracted with the help of saliency-based spatial attention and channel attention schemes. In addition, a saliency map serves as a guideline for predicting a patch weight map in order to afford stable training of end-to-end optimization for the JND-SalCAR. To the best of our knowledge, our work presents the first HVS-inspired trainable FR-IQA network that considers both visual saliency and the JND characteristics of the HVS. When the visual saliency map and the JND probability map are explicitly given as priors, they can be usefully combined to predict IQA scores rated by humans more precisely, eventually leading to performance improvements and faster convergence. The experimental results show that the proposed JND-SalCAR significantly outperforms all recent state-of-the-art FR-IQA methods on large IQA datasets in terms of the Spearman rank order coefficient (SRCC) and the Pearson linear correlation coefficient (PLCC).
引用
收藏
页码:2602 / 2616
页数:15
相关论文
共 6 条
  • [1] A new full-reference image quality metric based on just noticeable difference
    Toprak, Sevil
    Yalman, Yildiray
    COMPUTER STANDARDS & INTERFACES, 2017, 50 : 18 - 25
  • [2] A weighted full-reference image quality assessment based on visual saliency
    Wen, Yang
    Li, Ying
    Zhang, Xiaohua
    Shi, Wuzhen
    Wang, Lin
    Chen, Jiawei
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2017, 43 : 119 - 126
  • [3] Unifying Dual-Attention and Siamese Transformer Network for Full-Reference Image Quality Assessment
    Tang, Zhenjun
    Chen, Zhiyuan
    Li, Zhixin
    Zhong, Bineng
    Zhang, Xianquan
    ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2023, 19 (06)
  • [4] Neural Network-Based Full-Reference Image Quality Assessment
    Bosse, Sebastian
    Maniry, Dominique
    Mueller, Klaus-Robert
    Wiegand, Thomas
    Samek, Wojciech
    2016 PICTURE CODING SYMPOSIUM (PCS), 2016,
  • [5] Full-Reference Image Quality Assessment Based on Grunwald-Letnikov Derivative, Image Gradients, and Visual Saliency
    Varga, Domonkos
    ELECTRONICS, 2022, 11 (04)
  • [6] Full-Reference Image Quality Assessment Based on Multi-Channel Visual Information Fusion
    Jiang, Benchi
    Bian, Shilei
    Shi, Chenyang
    Wu, Lulu
    APPLIED SCIENCES-BASEL, 2023, 13 (15):