MPEG-4 AVC saliency map computation

被引:0
作者
Ammar, M. [2 ]
Mitrea, M. [1 ,2 ]
Hasnaoui, M. [2 ]
机构
[1] 9 Rue Charles Fourier, F-91011 Evry, France
[2] Telecom SudParis, Inst Mines Telecom, ARTEMIS Dept, UMR 8145, F-91011 Evry, France
来源
HUMAN VISION AND ELECTRONIC IMAGING XIX | 2014年 / 9014卷
关键词
Saliency map; MPEG-4 AVC saliency map; HVS; watermarking; MPEG-4 AVC stream; MODEL;
D O I
10.1117/12.2042450
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
A saliency map provides information about the regions inside some visual content (image, video,...) at which a human observer will spontaneously look at. For saliency maps computation, current research studies consider the uncompressed (pixel) representation of the visual content and extract various types of information (intensity, color, orientation, motion energy) which are then fusioned. This paper goes one step further and computes the saliency map directly from the MPEG-4 AVC stream syntax elements with minimal decoding operations. In this respect, an a-priori in-depth study on the MPEG-4 AVC syntax elements is first carried out so as to identify the entities appealing the visual attention. Secondly, the MPEG-4 AVC reference software is completed with software tools allowing the parsing of these elements and their subsequent usage in objective benchmarking experiments. This way, it is demonstrated that an MPEG-4 saliency map can be given by a combination of static saliency and motion maps. This saliency map is experimentally validated under a robust watermarking framework. When included in an m-QIM (multiple symbols Quantization Index Modulation) insertion method, PSNR average gains of 2.43 dB, 2.15dB, and 2.37 dB are obtained for data payload of 10, 20 and 30 watermarked blocks per I frame, i.e. about 30, 60, and 90 bits/second, respectively. These quantitative results are obtained out of processing 2 hours of heterogeneous video content.
引用
收藏
页数:10
相关论文
共 10 条
  • [1] Watermarking as communications with side information
    Cox, IJ
    Miller, ML
    McKellips, AL
    [J]. PROCEEDINGS OF THE IEEE, 1999, 87 (07) : 1127 - 1141
  • [2] Dongjian He, 2010, 2010 International Conference on Computer and Communication Technologies in Agriculture Engineering (CCTAE 2010), P323, DOI 10.1109/CCTAE.2010.5544608
  • [3] Predictive Saliency Maps for Surveillance Videos
    Guraya, Fahad Fazal Elahi
    Cheikh, Faouzi Alaya
    Tremeau, Alain
    Tong, Yubing
    Konik, Hubert
    [J]. PROCEEDINGS OF THE NINTH INTERNATIONAL SYMPOSIUM ON DISTRIBUTED COMPUTING AND APPLICATIONS TO BUSINESS, ENGINEERING AND SCIENCE (DCABES 2010), 2010, : 508 - 513
  • [4] Hasnaoui M., 2013, SINGAL PROCESSING IM, V29, P107
  • [5] Hasnaoui M., 2011, P SPIE, V7881
  • [6] Iain E., 2003, H 264 MPEG 4 VIDEO C
  • [7] Automatic foveation for video compression using a neurobiological model of visual attention
    Itti, L
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2004, 13 (10) : 1304 - 1318
  • [8] A model of saliency-based visual attention for rapid scene analysis
    Itti, L
    Koch, C
    Niebur, E
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 1998, 20 (11) : 1254 - 1259
  • [9] A coherent computational approach to model bottom-up visual attention
    Le Meur, O
    Le Callet, P
    Barba, D
    Thoreau, D
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2006, 28 (05) : 802 - 817
  • [10] Zhi L., 2009, P IEEE ACIS INT C CO, P568