3D visual saliency and convolutional neural network for blind mesh quality assessment

被引:22
|
作者
Abouelaziz, Ilyass [1 ]
Chetouani, Aladine [2 ]
El Hassouni, Mohammed [1 ,3 ]
Latecki, Longin Jan [4 ]
Cherifi, Hocine [5 ]
机构
[1] Mohammed V Univ Rabat, Fac Sci, LRIT, URAC 29, BP 1014 RP, Rabat, Morocco
[2] Univ Orleans, PRISME Lab, Orleans, France
[3] Mohammed V Univ Rabat, FLSHR, Rabat, Morocco
[4] Temple Univ, Dept Comp & Informat Sci, Philadelphia, PA 19122 USA
[5] Univ Burgundy, LE2I, UMR 6306, CNRS, Dijon, France
来源
NEURAL COMPUTING & APPLICATIONS | 2020年 / 32卷 / 21期
关键词
Mesh visual quality assessment; Mean opinion score; Mesh visual saliency; Convolutional neural network; METRICS; ERROR; COMPRESSION; MODEL;
D O I
10.1007/s00521-019-04521-1
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
A number of full reference and reduced reference methods have been proposed in order to estimate the perceived visual quality of 3D meshes. However, in most practical situations, there is a limited access to the information related to the reference and the distortion type. For these reasons, the development of a no-reference mesh visual quality (MVQ) approach is a critical issue, and more emphasis needs to be devoted to blind methods. In this work, we propose a no-reference convolutional neural network (CNN) framework to estimate the perceived visual quality of 3D meshes. The method is called SCNN-BMQA (3D visual saliency and CNN for blind mesh quality assessment). The main contribution is the usage of a CNN and 3D visual saliency to estimate the perceived visual quality of distorted meshes. To do so, the CNN architecture is fed by small patches selected carefully according to their level of saliency. First, the visual saliency of the 3D mesh is computed. Afterward, we render 2D projections from the 3D mesh and its corresponding 3D saliency map. Then the obtained views are split into 2D small patches that pass through a saliency filter in order to select the most relevant patches. Finally, a CNN is used for the feature learning and the quality score estimation. Extensive experiments are conducted on four prominent MVQ assessment databases, including several tests to study the effect of the CNN parameters, the effect of visual saliency and comparison with existing methods. Results show that the trained CNN achieves good rates in terms of correlation with human judgment and outperforms the most effective state-of-the-art methods.
引用
收藏
页码:16589 / 16603
页数:15
相关论文
共 50 条
  • [41] Pair-Comparing Based Convolutional Neural Network for Blind Image Quality Assessment
    Qin, Xue
    Xiang, Tao
    Yang, Ying
    Liao, Xiaofeng
    ADVANCES IN NEURAL NETWORKS - ISNN 2019, PT II, 2019, 11555 : 460 - 468
  • [42] 3D convolutional neural network for object recognition: a review
    Singh, Rahul Dev
    Mittal, Ajay
    Bhatia, Rajesh K.
    MULTIMEDIA TOOLS AND APPLICATIONS, 2019, 78 (12) : 15951 - 15995
  • [43] Subjective Visual Quality Assessment of Immersive 3D Media Compressed by Open-Source Static 3D Mesh Codecs
    Christaki, Kyriaki
    Christakis, Emmanouil
    Drakoulis, Petros
    Doumanoglou, Alexandros
    Zioulis, Nikolaos
    Zarpalas, Dimitrios
    Daras, Petros
    MULTIMEDIA MODELING (MMM 2019), PT I, 2019, 11295 : 80 - 91
  • [44] Subjective and Objective Visual Quality Assessment of Textured 3D Meshes
    Guo, Jinjiang
    Vidal, Vincent
    Cheng, Irene
    Basu, Anup
    Baskurt, Atilla
    Lavoue, Guillaume
    ACM TRANSACTIONS ON APPLIED PERCEPTION, 2017, 14 (02)
  • [45] ECG Quality Assessment Using 1D-Convolutional Neural Network
    Zhou, Xue
    Zhu, Xin
    Nakamura, Keijiro
    Mahito, Noro
    PROCEEDINGS OF 2018 14TH IEEE INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING (ICSP), 2018, : 780 - 784
  • [46] Towards 3D Colored Mesh Saliency: Database and Benchmarks
    Ding, Xiaoying
    Chen, Zhao
    Lin, Weisi
    Chen, Zhenzhong
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 3580 - 3591
  • [47] Applying deep convolutional neural network with 3D reality mesh model for water tank crack detection and evaluation
    Wu, Z. Y.
    Kalfarisi, R.
    Kouyoumdjian, F.
    Taelman, C.
    URBAN WATER JOURNAL, 2020, 17 (08) : 682 - 695
  • [48] Learning Human Motion Intention with 3D Convolutional Neural Network
    Owoyemi, Joshua
    Hashimoto, Koichi
    2017 IEEE INTERNATIONAL CONFERENCE ON MECHATRONICS AND AUTOMATION (ICMA), 2017, : 1810 - 1815
  • [49] Explainable 3D convolutional neural network using GMM encoding
    Stano, Martin
    Benesova, Wanda
    Martak, Lukas S.
    TWELFTH INTERNATIONAL CONFERENCE ON MACHINE VISION (ICMV 2019), 2020, 11433
  • [50] 3D Human Motion Synthesis Based on Convolutional Neural Network
    Zhou, Dongsheng
    Feng, Xinzhu
    Yi, Pengfei
    Yang, Xin
    Zhang, Qiang
    Wei, Xiaopeng
    Yang, Deyun
    IEEE ACCESS, 2019, 7 : 66325 - 66335