No-Reference Video Quality Assessment Metric Using Spatiotemporal Features Through LSTM

被引:0
|
作者
Kwong, Ngai-Wing [1 ]
Tsang, Sik-Ho [2 ]
Chan, Yui-Lam [1 ]
Lun, Daniel Pak-Kong [1 ,2 ]
Lee, Tsz-Kwan [3 ]
机构
[1] Hong Kong Polytech Univ, Dept Elect & Informat Engn, Hong Kong, Peoples R China
[2] Ctr Adv Reliabil & Safety Ltd CAiRS, Hong Kong Sci Pk, Hong Kong, Peoples R China
[3] Deakin Univ, Sch Informat Technol, Deakin, Australia
关键词
video quality assessment; no reference; long short-term memory; spatiotemporal; pre-padding; masking layer;
D O I
10.1117/12.2590406
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Nowadays, a precise video quality assessment (VQA) model is essential to maintain the quality of service (QoS). However, most existing VQA metrics are designed for specific purposes and ignore the spatiotemporal features of nature video. This paper proposes a novel general-purpose no-reference (NR) VQA metric adopting Long Short-Term Memory (LSTM) modules with the masking layer and pre-padding strategy, namely VQA-LSTM, to solve the above issues. First, we divide the distorted video into frames and extract some significant but also universal spatial and temporal features that could effectively reflect the quality of frames. Second, the data preprocessing stage and pre-padding strategy are used to process data to ease the training for our VQA-LSTM. Finally, a three-layer LSTM model incorporated with masking layer is designed to learn the sequence of spatial features as spatiotemporal features and learn the sequence of temporal features as the gradient of temporal features to evaluate the quality of videos. Two widely used VQA database, MCL-V and LIVE, are tested to prove the robustness of our VQA-LSTM, and the experimental results show that our VQA-LSTM has a better correlation with human perception than some state-of-the-art approaches.
引用
收藏
页数:6
相关论文
共 50 条
  • [21] A Fast and Efficient No-Reference Video Quality Assessment Algorithm Using Video Action Recognition Features
    Suresh, N.
    Mylavarapu, Pavan Manesh
    Mahankali, Naga Sailaja
    Channappayya, Sumohana S.
    2022 NATIONAL CONFERENCE ON COMMUNICATIONS (NCC), 2022, : 402 - 406
  • [22] No-Reference Video Quality Assessment Using Local Structural and Quality-Aware Deep Features
    Vishwakarma, Anish Kumar
    Bhurchandi, Kishor M.
    IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2023, 72
  • [23] NO-REFERENCE VIDEO QUALITY ASSESSMENT USING MPEG ANALYSIS
    Sogaard, Jacob
    Forchhammer, Soren
    Korhonen, Jari
    2013 PICTURE CODING SYMPOSIUM (PCS), 2013, : 161 - 164
  • [24] No-Reference Video Quality Assessment Using Codec Analysis
    Sogaard, Jacob
    Forchhammer, Soren
    Korhonen, Jari
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2015, 25 (10) : 1637 - 1650
  • [25] QMET: A New Quality Assessment Metric for No-Reference Video Coding by Using Human Eye Traversal
    Podder, Pallab Kanti
    Paul, Manoranjan
    Murshed, Manzur
    PROCEEDINGS OF THE 2016 INTERNATIONAL CONFERENCE ON IMAGE AND VISION COMPUTING NEW ZEALAND (IVCNZ), 2016, : 140 - 145
  • [26] No-Reference Video Quality Metric for Streaming Service Using DASH Standard
    Rodriguez, Demostenes Zegarra
    Rosa, Renata Lopes
    Bressan, Graca
    2015 IEEE INTERNATIONAL CONFERENCE ON CONSUMER ELECTRONICS (ICCE), 2015, : 106 - 107
  • [27] No-Reference Video Quality Assessment Using the Temporal Statistics of Global and Local Image Features
    Varga, Domonkos
    SENSORS, 2022, 22 (24)
  • [28] No-Reference Video Shakiness Quality Assessment
    Cui, Zhaoxiong
    Jiang, Tingting
    COMPUTER VISION - ACCV 2016, PT V, 2017, 10115 : 396 - 411
  • [29] COME for No-Reference Video Quality Assessment
    Wang, Chunfeng
    Su, Li
    Zhang, Weigang
    IEEE 1ST CONFERENCE ON MULTIMEDIA INFORMATION PROCESSING AND RETRIEVAL (MIPR 2018), 2018, : 232 - 237
  • [30] Predictive no-reference assessment of video quality
    Vega, Maria Torres
    Mocanu, Decebal Constantin
    Stavrou, Stavros
    Liotta, Antonio
    SIGNAL PROCESSING-IMAGE COMMUNICATION, 2017, 52 : 20 - 32