An egocentric video and eye-tracking dataset for visual search in convenience stores

被引:0
作者
Wang, Yinan [1 ]
Panchadsaram, Sansitha [1 ]
Sherkati, Rezvan [1 ]
Clark, James J. [1 ]
机构
[1] McGill Univ, Dept Elect & Comp Engn, Montreal, PQ, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
Attention; Saliency; Eye-tracking; Egocentric; SALIENCY; IMAGE; ATTENTION; OBJECTS; PEOPLE; MODEL;
D O I
10.1016/j.cviu.2024.104129
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We introduce an egocentric video and eye-tracking dataset, comprised of 108 first-person videos of 36 shoppers searching for three different products (orange juice, KitKat chocolate bars, and canned tuna) in a convenience store, along with the frame-centered eye fixation locations for each video frame. The dataset also includes demographic information about each participant in the form of an 11-question survey. The paper describes two applications using the dataset - an analysis of eye fixations during search in the store, and a training of a clustered saliency model for predicting saliency of viewers engaged in product search in the store. The fixation analysis shows that fixation duration statistics are very similar to those found in image and video viewing, suggesting that similar visual processing is employed during search in 3D environments and during viewing of imagery on computer screens. A clustering technique was applied to the questionnaire data, which resulted in two clusters being detected. Based on these clusters, personalized saliency prediction models were trained on the store fixation data, which provided improved performance in prediction saliency on the store video data compared to state-of-the art universal saliency prediction methods.
引用
收藏
页数:18
相关论文
共 74 条
[1]   When will you do what? - Anticipating Temporal Occurrences of Activities [J].
Abu Farha, Yazan ;
Richard, Alexander ;
Gall, Juergen .
2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, :5343-5352
[2]   Fast unfolding of communities in large networks [J].
Blondel, Vincent D. ;
Guillaume, Jean-Loup ;
Lambiotte, Renaud ;
Lefebvre, Etienne .
JOURNAL OF STATISTICAL MECHANICS-THEORY AND EXPERIMENT, 2008,
[3]  
Borji A, 2015, Arxiv, DOI [arXiv:1505.03581, 10.48550/arXiv.1505.03581]
[4]   The Architecture of Object-Based Attention [J].
Cavanagh, Patrick ;
Caplovitz, Gideon P. ;
Lytchenko, Taissa K. ;
Maechler, Marvin R. ;
Tse, Peter U. ;
Sheinberg, David L. .
PSYCHONOMIC BULLETIN & REVIEW, 2023, 30 (05) :1643-1667
[5]   Spatio-temporal distribution characteristics and influencing factors of COVID-19 in China [J].
Chen, Youliang ;
Li, Qun ;
Karimian, Hamed ;
Chen, Xunjun ;
Li, Xiaoming .
SCIENTIFIC REPORTS, 2021, 11 (01)
[6]  
Cornia M, 2016, INT C PATT RECOG, P3488, DOI 10.1109/ICPR.2016.7900174
[7]   Scaling Egocentric Vision: The EPIC-KITCHENS Dataset [J].
Damen, Dima ;
Doughty, Hazel ;
Farinella, Giovanni Maria ;
Fidler, Sanja ;
Furnari, Antonino ;
Kazakos, Evangelos ;
Moltisanti, Davide ;
Munro, Jonathan ;
Perrett, Toby ;
Price, Will ;
Wray, Michael .
COMPUTER VISION - ECCV 2018, PT IV, 2018, 11208 :753-771
[8]  
Darkhalil Ahmad, 2022, ADV NEUR IN
[9]   Individual differences in visual salience vary along semantic dimensions [J].
de Haas, Benjamin ;
Iakovidis, Alexios L. ;
Schwarzkopf, D. Samuel ;
Gegenfurtner, Karl R. .
PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2019, 116 (24) :11687-11692
[10]   Summarization of Egocentric Videos: A Comprehensive Survey [J].
del Molino, Ana Garcia ;
Tan, Cheston ;
Lim, Joo-Hwee ;
Tan, Ah-Hwee .
IEEE TRANSACTIONS ON HUMAN-MACHINE SYSTEMS, 2017, 47 (01) :65-76