Blind Image Quality Assessment via Cross-View Consistency

被引:57
作者
Zhu, Yucheng [1 ,2 ,3 ]
Li, Yunhao [1 ,2 ]
Sun, Wei [1 ,2 ]
Min, Xiongkuo [1 ,2 ]
Zhai, Guangtao [1 ,2 ]
Yang, Xiaokang [1 ,2 ]
机构
[1] Shanghai Jiao Tong Univ, Inst Image Commun & Network Engn, Shanghai 200240, Peoples R China
[2] Shanghai Jiao Tong Univ, Moe Key Lab Artificial Intelligence, Shanghai 200240, Peoples R China
[3] USC SJTU Inst Cultural & Creat Ind, Shanghai 200240, Peoples R China
基金
中国博士后科学基金; 国家重点研发计划; 中国国家自然科学基金;
关键词
Image quality; Feature extraction; Distortion; Image color analysis; Data mining; Transformers; Task analysis; Image quality assessment; authentic distortion; self-supervised learning; transformer; deep learning; LEVEL;
D O I
10.1109/TMM.2022.3224319
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Image quality assessment (IQA) is very important for both end-users and service-providers since a high-quality image can significantly improve the user's quality of experience (QoE). Most existing blind image quality assessment (BIQA) models were developed for synthetically distorted images, however, they perform poorly on in-the-wild images, which are widely existed in various practical applications. In this paper, a BIQA model is proposed that consists of a desirable self-supervised feature learning approach to mitigate the data shortage problem and learn comprehensive feature representations, and a self-attention-based feature fusion module to introduce self-attention mechanism. We develop the image quality assessment model under the framework of contrastive learning with multi views. Since human visual system perceives signals through multiple channels, the most important visual information should exist among all views of the channels. So we design the cross-view consistent information mining (CVC-IM) module to extract compact mutual information between different views. Color information and pseudo-reference image (PRI) of different distortion types are employed to formulate rich feature embeddings and preserve the quality-aware fidelity of learned representations. We employ the Transformer as the self-attention-based architecture to integrate feature embeddings. Extensive experiments show that our model achieves remarkable image quality assessment results on in-the-wild IQA datasets.
引用
收藏
页码:7607 / 7620
页数:14
相关论文
共 60 条
[1]   NTIRE 2017 Challenge on Single Image Super-Resolution: Dataset and Study [J].
Agustsson, Eirikur ;
Timofte, Radu .
2017 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW), 2017, :1122-1131
[2]  
Ali M.A., 2012, Vision in Vertebrates
[3]  
Antkowiak J., 2000, document COM9-80-E
[4]   Deep Neural Networks for No-Reference and Full-Reference Image Quality Assessment [J].
Bosse, Sebastian ;
Maniry, Dominique ;
Mueller, Klaus-Robert ;
Wiegand, Thomas ;
Samek, Wojciech .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2018, 27 (01) :206-219
[5]   Blind quality assessment of authentically distorted images [J].
Celona, Luigi ;
Schettini, Raimondo .
JOURNAL OF THE OPTICAL SOCIETY OF AMERICA A-OPTICS IMAGE SCIENCE AND VISION, 2022, 39 (06) :B1-B10
[6]   Unsupervised Curriculum Domain Adaptation for No-Reference Video Quality Assessment [J].
Chen, Pengfei ;
Li, Leida ;
Wu, Jinjian ;
Dong, Weisheng ;
Shi, Guangming .
2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, :5158-5167
[7]   Perceptual Image Quality Assessment with Transformers [J].
Cheon, Manri ;
Yoon, Sung-Jun ;
Kang, Byungyeon ;
Lee, Junwoo .
2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2021, 2021, :433-442
[8]   How prediction errors shape perception, attention, and motivation [J].
den Ouden, Hanneke E. M. ;
Kok, Peter ;
de Lange, Floris P. .
FRONTIERS IN PSYCHOLOGY, 2012, 3
[9]  
Deng J, 2009, PROC CVPR IEEE, P248, DOI 10.1109/CVPRW.2009.5206848
[10]  
Devlin J, 2019, 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, P4171