Crowdsourced Saliency for Mining Robotically Gathered 3D Maps Using Multitouch Interaction on Smartphones and Tablets

被引:0
作者
Johnson-Roberson, Matthew [1 ]
Bryson, Mitch [2 ]
Douillard, Bertrand [3 ]
Pizarro, Oscar [2 ]
Williams, Stefan B. [2 ]
机构
[1] Univ Michigan, Dept Naval Architecture & Marine Engn, Ann Arbor, MI 48109 USA
[2] Univ Sydney, Australian Ctr Field Robot, Sydney, NSW 2006, Australia
[3] CALTECH, NASA, Jet Prop Lab, Pasadena, CA 91125 USA
来源
2014 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA) | 2014年
关键词
ATTENTION;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper presents a system for crowdsourcing saliency interest points for robotically gathered 3D maps rendered on smartphones and tablets. An app was created that is capable of interactively rendering 3D reconstructions gathered with an Autonomous Underwater Vehicle. Through hundreds of thousands of logged user interactions with the models we attempt to data-mine salient interest points. To this end we propose two models for calculating saliency from human interaction with the data. The first uses the view frustum of the camera to track the amount of time points are on screen. The second treats the camera's path as a time series and uses a Hidden Markov model to learn the classification of salient and non-salient points. To provide a comparison to existing techniques, several traditional visual saliency approaches are applied to orthographic views of the models' photo-texturing. The results of all approaches are validated with human attention ground truth gathered using a remote gaze-tracking system that recorded the locations of the person's attention while exploring the models.
引用
收藏
页码:6032 / 6039
页数:8
相关论文
共 25 条
[1]  
[Anonymous], 2007, PROC IEEE C COMPUT V, DOI 10.1109/CVPR.2007.383267
[2]  
[Anonymous], P IEEE CVPR 2012 EG
[3]  
Baccot B., 2008, 2008 Ninth International Workshop on Image Analysis for Multimedia Interactive Services (WIAMIS), P122, DOI 10.1109/WIAMIS.2008.57
[4]  
Carlier Axel, 2010, P 18 ACM INT C MULT, P201
[5]   Multimodal Semantics Extraction from User-Generated Videos [J].
Cricri, Francesco ;
Dabov, Kostadin ;
Roininen, Mikko J. ;
Mate, Sujeet ;
Curcio, Igor D. D. ;
Gabbouj, Moncef .
ADVANCES IN MULTIMEDIA, 2012, 2012
[6]   A hierarchical neural system with attentional top-down enhancement of the spatial resolution for object recognition [J].
Deco, G ;
Schürmann, B .
VISION RESEARCH, 2000, 40 (20) :2845-2859
[7]  
Harel J., 2006, Graph-Based Visual Saliency, V19, DOI DOI 10.7551/MITPRESS/7503.003.0073
[8]   Image Signature: Highlighting Sparse Salient Regions [J].
Hou, Xiaodi ;
Harel, Jonathan ;
Koch, Christof .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2012, 34 (01) :194-201
[9]   A model of saliency-based visual attention for rapid scene analysis [J].
Itti, L ;
Koch, C ;
Niebur, E .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 1998, 20 (11) :1254-1259
[10]   Generation and Visualization of Large-Scale Three-Dimensional Reconstructions from Underwater Robotic Surveys [J].
Johnson-Roberson, Matthew ;
Pizarro, Oscar ;
Williams, Stefan B. ;
Mahon, Ian .
JOURNAL OF FIELD ROBOTICS, 2010, 27 (01) :21-51