Visual saliency detection: An information theoretic algorithm combined long-term with short-term features

被引:0
作者
机构
[1] Institute of Automation, Northwestern Polytechnical University
来源
Qian, X.-L. (qxlzengli@gmail.com) | 1636年 / Science Press卷 / 35期
关键词
Information entropy; Long-term features; Pattern recognition; Shannon self-information; Short-term features; Visual saliency detection;
D O I
10.3724/SP.J.1146.2012.01251
中图分类号
学科分类号
摘要
In order for removing the drawback of the traditional visual saliency detection methods which solely used the information of current viewing image or prior knowledge, this paper proposes an information theoretic algorithm to combine the long-term features which imply the prior knowledge with short-term features which imply the information of current viewing image. Firstly, a long-term sparse dictionary and short-term sparse dictionary are trained using the eye-tracking data and current viewing image, respectively. Their corresponding sparse codes are regarded as the long-term and short-term features, respectively. Secondly, to reduce the problem of existing methods which derivated features on the entire image or a local neighborhood with the fixed size, an information entropy based the estimation method of probability distribution of features is proposed. This method can infer an optimal size of region adaptively according to the characteristics of the current viewing image for the calculation of probability of the appearance of long-tern and short-term features. Finally, the saliency map is formulated by Shannon self-information. The subjective and quantitative comparisons with 8 state-of-the-art methods on publicly available eye-tracking databases demonstrate the effectiveness of the proposed method.
引用
收藏
页码:1636 / 1643
页数:7
相关论文
共 27 条
  • [1] Rutishauser U., Walther D., Koch C., Et al., Is bottom-up attention useful for object recognition?, IEEE Conference on Computer Vision and Pattern Recognition, pp. 37-44, (2004)
  • [2] Han J.W., Ngan K.N., Li M.J., Et al., Unsupervised extraction of visual attention objects in color images, IEEE Transactions on Circuits and Systems for Video Technology, 16, 1, pp. 141-145, (2006)
  • [3] Wang Z., Lu L.G., Bovik A.C., Foveation scalable video coding with automatic fixation selection, IEEE Transactions on Image Processing, 12, 2, pp. 243-254, (2003)
  • [4] Yang J., Yang M.H., Top-down visual saliency via joint CRF and dictionary learning, IEEE Conference on Computer Vision and Pattern Recognition, pp. 2296-2303, (2012)
  • [5] Koch C., Ullman S., Shifts in selective visual attention: Towards the underlying neural circuitry, Human Neurobiology, 4, 4, pp. 219-227, (1985)
  • [6] Itti L., Koch C., Niebur E., A model of saliency-based visual attention for rapid scene analysis, IEEE Transactions on Pattern Analysis and Machine Intelligence, 20, 11, pp. 1254-1259, (1998)
  • [7] Zhang J., Shen L.-S., Gao J.-J., Region of interest detection based on visual attention model and evolutionary programming, Journal of Electronics and Information Technology, 31, 7, pp. 1646-1652, (2009)
  • [8] Murray N., Vanrell M., Otazu X., Et al., Saliency estimation using a non-parametric low-level vision model, IEEE Conference on Computer Vision and Pattern Recognition, pp. 433-440, (2011)
  • [9] Harel J., Koch C., Perona P., Graph-based visual saliency, Advances in Neural Information Processing Systems, pp. 545-552, (2007)
  • [10] Avraham T., Lindenbaum M., Esaliency (extended saliency): Meaningful attention using stochastic image modeling, IEEE Transactions on Pattern Analysis and Machine Intelligence, 32, 4, pp. 693-708, (2010)