3D Layout encoding network for spatial-aware 3D saliency modelling

被引:2
|
作者
Yuan, Jing [1 ]
Cao, Yang [2 ]
Kang, Yu [2 ]
Song, Weiguo [1 ]
Yin, Zhongcheng [2 ]
Ba, Rui [1 ]
Ma, Qing [1 ]
机构
[1] Univ Sci & Technol China, State Key Lab Fire Sci, Hefei, Anhui, Peoples R China
[2] Univ Sci & Technol China, Sch Informat Sci & Technol, Dept Automat, Hefei, Anhui, Peoples R China
基金
中国国家自然科学基金;
关键词
image sensors; image fusion; object detection; image colour analysis; feature extraction; popular 3D multimedia applications; existing 3D devices; low quality; holes; predictions; single depth images; deep layout features; spatial-aware saliency prediction; coarse depth-induced saliency cues; depth details; high-quality RGB image; low-level; final prediction; spatial layout; saliency modelling results; OBJECT DETECTION; VISUAL-ATTENTION;
D O I
10.1049/iet-cvi.2018.5591
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Three-dimensional (3D) [red, green and blue (RGB) + depth] saliency modelling can help with popular 3D multimedia applications. However, depth images produced from existing 3D devices are often with low quality, e.g. containing noises and holes. In this study, rather than relying on features or predictions directly derived from single depth images, the authors propose to encode deep layout features to facilitate the spatial-aware saliency prediction. Specifically, they first generate coarse depth-induced saliency cues which are careless of depth details. Then, to leverage the information of the high-quality RGB image, they embed both low-level and high-level RGB deep features to refine the final prediction. In this way, they take both bottom-up and top-down cues together with spatial layout into account and achieve better saliency modelling results. Experiments on five public datasets show the superiority of the proposed method.
引用
收藏
页码:480 / 488
页数:9
相关论文
共 50 条
  • [31] Saliency-Guided 3D Head Pose Estimation on 3D Expression Models
    Liu, Peng
    Reale, Michael
    Zhang, Xing
    Yin, Lijun
    ICMI'13: PROCEEDINGS OF THE 2013 ACM INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION, 2013, : 75 - 78
  • [32] 3D MORPHOLOGICAL MODELLING OF A RANDOM FIBROUS NETWORK
    Peyrega, Charles
    Jeulin, Dominique
    Delisee, Christine
    Malvestio, Jerome
    ECS10: THE10TH EUROPEAN CONGRESS OF STEREOLOGY AND IMAGE ANALYSIS, 2009, : 125 - +
  • [33] Adaptive Multi-level Saliency Network in 3D Generation
    Tang, Zhongliang
    2018 IEEE 3RD INTERNATIONAL CONFERENCE ON IMAGE, VISION AND COMPUTING (ICIVC), 2018, : 906 - 910
  • [34] MVSGAN: Spatial-Aware Multi-View CMR Fusion for Accurate 3D Left Ventricular Myocardium Segmentation
    Qi, Xiaoming
    He, Yuting
    Yang, Guanyu
    Chen, Yang
    Yang, Jian
    Liu, Wangyag
    Zhu, Yinsu
    Xu, Yi
    Shu, Huazhong
    Li, Shuo
    IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2022, 26 (05) : 2264 - 2275
  • [35] Drawing and 3D modelling
    Docci, Mario
    DISEGNARE IDEE IMMAGINI-IDEAS IMAGES, 2015, 26 (51): : 5 - 6
  • [36] 3D modelling and recognition
    Rodrigues, M.
    Robinson, A.
    Alboul, L.
    Brink, W.
    WSEAS Transactions on Information Science and Applications, 2006, 3 (11): : 2118 - 2122
  • [37] 3D MOTION IN VISUAL SALIENCY MODELING
    Wan, Pengfei
    Feng, Yunlong
    Cheung, Gene
    Bajic, Ivan V.
    Au, Oscar C.
    Ji, Yusheng
    2013 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2013, : 1831 - 1835
  • [38] Learning to Predict 3D Mesh Saliency
    ALfarasani, Dalia A.
    Sweetman, Thomas
    Lai, Yu-Kun
    Rosin, Paul L.
    2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 4023 - 4029
  • [39] Image retargeting with a 3D saliency model
    Chen, Yanxiang
    Pan, Yifei
    Song, Minglong
    Wang, Meng
    SIGNAL PROCESSING, 2015, 112 : 53 - 63
  • [40] Saliency Regions for 3D Mesh Abstraction
    Yang, Yu-Bin
    Lu, Tong
    Lin, Jin-Jie
    ADVANCES IN MULTIMEDIA INFORMATION PROCESSING - PCM 2009, 2009, 5879 : 292 - 299