Saliency Sandbox Bottom-up Saliency Framework

被引:1
作者
Geisler, David [1 ]
Fuhl, Wolfgang [1 ]
Santini, Thiago [1 ]
Kasneci, Enkelejda [1 ]
机构
[1] Univ Tubingen, Percept Engn, Tubingen, Germany
来源
PROCEEDINGS OF THE 12TH INTERNATIONAL JOINT CONFERENCE ON COMPUTER VISION, IMAGING AND COMPUTER GRAPHICS THEORY AND APPLICATIONS (VISIGRAPP 2017), VOL 4 | 2017年
关键词
Saliency Sandbox; Feature Maps; Attention Maps; Saliency Maps; Bottom Up;
D O I
10.5220/0006272306570664
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Saliency maps are used to predict the visual stimulus raised from a certain region in a scene. Most approaches to calculate the saliency in a scene can be divided into three consecutive steps: extraction of feature maps, calculation of activation maps, and the combination of activation maps. In the past two decades, several new saliency estimation approaches have emerged. However, most of these approaches are not freely available as source code, thus requiring researchers and application developers to reimplement them. Moreover, others are freely available but use different platforms for their implementation. As a result, employing, evaluating, and combining existing approaches is time consuming, costly, and even error-prone (e.g., when reimplementation is required). In this paper, we introduce the Saliency Sandbox, a framework for the fast implementation and prototyping of saliency maps, which employs a flexible architecture that allows designing new saliency maps by combining existing and new approaches such as Itti & Koch, GBVS, Boolean Maps and many more. The Saliency Sandbox comes with a large set of implemented feature extractors as well as some of the most popular activation approaches. The framework core is written in C++; nonetheless, interfaces for Matlab and Simulink allow for fast prototyping and integration of already existing implementations. Our source code is available at: www.ti.uni-tuebingen.de/perception.
引用
收藏
页码:657 / 664
页数:8
相关论文
共 21 条
  • [1] Achanta R, 2009, PROC CVPR IEEE, P1597, DOI 10.1109/CVPRW.2009.5206596
  • [2] [Anonymous], 2007, PROC IEEE C COMPUT V, DOI 10.1109/CVPR.2007.383267
  • [3] [Anonymous], 2011, ITU-R Recommendation BT.601: Studio encoding parameters of digital television for standard 4:3 and wide screen 16:9 aspect ratios
  • [4] On the necessity of adaptive eye movement classification in conditionally automated driving scenarios
    Braunagel, Christian
    Geisler, David
    Stolzmann, Wolfgang
    Rosenstiel, Wolfgang
    Kasneci, Enkelejda
    [J]. 2016 ACM SYMPOSIUM ON EYE TRACKING RESEARCH & APPLICATIONS (ETRA 2016), 2016, : 19 - 26
  • [5] CHROMATIC MECHANISMS IN LATERAL GENICULATE-NUCLEUS OF MACAQUE
    DERRINGTON, AM
    KRAUSKOPF, J
    LENNIE, P
    [J]. JOURNAL OF PHYSIOLOGY-LONDON, 1984, 357 (DEC): : 241 - 265
  • [6] Godbehere AB, 2012, P AMER CONTR CONF, P4305
  • [7] Harel J., 2007, ADV NEURAL INFORM PR, P545, DOI DOI 10.7551/MITPRESS/7503.003.0073
  • [8] A saliency-based search mechanism for overt and covert shifts of visual attention
    Itti, L
    Koch, C
    [J]. VISION RESEARCH, 2000, 40 (10-12) : 1489 - 1506
  • [9] Itti L., 2004, NEUROMORPHIC ENG, V1, P10
  • [10] KaewTraKulPong P, 2002, VIDEO-BASED SURVEILLANCE SYSTEMS: COMPUTER VISION AND DISTRIBUTED PROCESSING, P135