Lightweight transformer and multi-head prediction network for no-reference image quality assessment

被引:3
作者
Tang, Zhenjun [1 ]
Chen, Yihua [1 ]
Chen, Zhiyuan [1 ]
Liang, Xiaoping [1 ]
Zhang, Xianquan [1 ]
机构
[1] Guangxi Normal Univ, Key Lab Educ Blockchain & Intelligent Technol, Minist Educ, Guilin 541004, Peoples R China
基金
中国国家自然科学基金;
关键词
Lightweight transformer; Multi-head prediction; Channel attention; Image quality assessment; NATURAL SCENE STATISTICS;
D O I
10.1007/s00521-023-09188-3
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
No-reference (NR) image quality assessment (IQA) is an important task of computer vision. Most NR-IQA methods via deep neural networks do not reach desirable IQA performance and have bulky models which make them difficult to be used in the practical scenarios. This paper proposes a lightweight transformer and multi-head prediction network for NR-IQA. The proposed method consists of two lightweight modules: feature extraction and multi-head prediction. The module of feature extraction exploits lightweight transformer blocks to learn features at different scales for measuring different image distortions. The module of multi-head prediction uses three weighted prediction blocks and an FC layer to aggregate the learned features for predicting image quality score. The weighted prediction block can measure the importance of different elements of input feature at the same scale. Since the importance of feature elements at the same scale and the importance of the features at different scales are both considered, the module of multi-head prediction can provide more accurate prediction results. Extensive experiments on the standard IQA datasets are conducted. The results show that the proposed method outperforms some baseline NR-IQA methods in IQA performance on the large image datasets. For the model complexity, the proposed method is also superior to several recent NR-IQA methods.
引用
收藏
页码:1947 / 1957
页数:11
相关论文
共 57 条
  • [1] Deep Neural Networks for No-Reference and Full-Reference Image Quality Assessment
    Bosse, Sebastian
    Maniry, Dominique
    Mueller, Klaus-Robert
    Wiegand, Thomas
    Samek, Wojciech
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2018, 27 (01) : 206 - 219
  • [2] A genetic programming-based convolutional neural network for image quality evaluations
    Chan, Kit Yan
    Lam, Hak-Keung
    Jiang, Huimin
    [J]. NEURAL COMPUTING & APPLICATIONS, 2022, 34 (18) : 15409 - 15427
  • [3] Chen YW, 2023, PROCEEDINGS OF 2023 7TH INTERNATIONAL CONFERENCE ON ELECTRONIC INFORMATION TECHNOLOGY AND COMPUTER ENGINEERING, EITCE 2023, P149, DOI [10.1007/978-981-19-9376-3_17, 10.1145/3650400.3650425]
  • [4] Multi-Level Feature Aggregation Network for Full-Reference Image Quality Assessment
    Chen, Zhiyuan
    Chen, Yihua
    Liang, Xiaoping
    Tang, Zhenjun
    [J]. 2022 IEEE 34TH INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE, ICTAI, 2022, : 861 - 866
  • [5] Generating Image Distortion Maps Using Convolutional Autoencoders With Application to No Reference Image Quality Assessment
    Dendi, Sathya Veera Reddy
    Dev, Chander
    Kothari, Narayan
    Channappayya, Sumohana S.
    [J]. IEEE SIGNAL PROCESSING LETTERS, 2019, 26 (01) : 89 - 93
  • [6] Dosovitskiy A., 2021, arXiv
  • [7] A survey of visual neural networks: current trends, challenges and opportunities
    Feng, Ping
    Tang, Zhenjun
    [J]. MULTIMEDIA SYSTEMS, 2023, 29 (02) : 693 - 724
  • [8] A No-Reference Objective Image Sharpness Metric Based on the Notion of Just Noticeable Blur (JNB)
    Ferzli, Rony
    Karam, Lina J.
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2009, 18 (04) : 717 - 728
  • [9] No-Reference Image Quality Assessment via Transformers, Relative Ranking, and Self-Consistency
    Golestaneh, S. Alireza
    Dadsetan, Saba
    Kitani, Kris M.
    [J]. 2022 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2022), 2022, : 3989 - 3999
  • [10] KonIQ-10k: An Ecologically Valid Database for Deep Learning of Blind Image Quality Assessment
    Hosu, Vlad
    Lin, Hanhe
    Sziranyi, Tamas
    Saupe, Dietmar
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 (29) : 4041 - 4056