Towards Predicting Reading Comprehension From Gaze Behavior

被引:20
作者
Ahn, Seoyoung [1 ]
Kelton, Conor [1 ]
Balasubramanian, Aruna [1 ]
Zelinsky, Gregory [1 ]
机构
[1] SUNY Stony Brook, Stony Brook, NY 11794 USA
来源
ETRA 2020 SHORT PAPERS: ACM SYMPOSIUM ON EYE TRACKING RESEARCH & APPLICATIONS | 2020年
基金
美国国家科学基金会;
关键词
Text comprehension prediction; Eye tracking; Machine learning; Reading dataset; EYE-MOVEMENTS;
D O I
10.1145/3379156.3391335
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
As readers of a language, we all agree to move our eyes in roughly the same way. Yet might there be hidden within this self-similar behavior subtle clues as to how a reader is understanding the material being read? Here we attempt to decode a reader's eye movements to predict their level of text comprehension and related states. Eye movements were recorded from 95 people reading 4 published SAT passages, each followed by corresponding SAT questions and self-evaluation questionnaires. A sequence of 21 fixation-location (x,y), fixation-duration, and pupil-size features were extracted from the reading behavior and input to two deep networks (CNN/RNN), which were used to predict the reader's comprehension level and other comprehension-related variables. The best overall comprehension prediction accuracy was 65% (cf. null accuracy = 54%) obtained by CNN. This prediction generalized well to fixations on new passages (64%) from the same readers, but did not generalize to fixations from new readers (41%), implying substantial individual differences in reading behavior. Our work is the first attempt to predict comprehension from fixations using deep networks, where we hope that our large reading dataset and our protocol for evaluation will benefit the development of new methods for predicting reading comprehension by decoding gaze behavior.
引用
收藏
页数:5
相关论文
共 16 条
  • [1] The impact of personality traits on users' information-seeking behavior
    Al-Samarraie, Hosam
    Eldenfria, Atef
    Dawoud, Husameddin
    [J]. INFORMATION PROCESSING & MANAGEMENT, 2017, 53 (01) : 237 - 247
  • [2] Al-Samarraie Hosam, 2018, COGNITION TECHNOLOGY, P1
  • [3] Augereau O, 2016, INT C PATT RECOG, P1285, DOI 10.1109/ICPR.2016.7899814
  • [4] Biedert R., 2012, P S EYE TRACK RES AP, P123, DOI DOI 10.1145/2168556.2168575
  • [5] Predicting task from eye movements: On the importance of spatial distribution, dynamics, and image features
    Boisvert, Jonathan F. G.
    Bruce, Neil D. B.
    [J]. NEUROCOMPUTING, 2016, 207 : 653 - 668
  • [6] Framewise phoneme classification with bidirectional LSTM and other neural network architectures
    Graves, A
    Schmidhuber, J
    [J]. NEURAL NETWORKS, 2005, 18 (5-6) : 602 - 610
  • [7] Human gaze control during real-world scene perception
    Henderson, JM
    [J]. TRENDS IN COGNITIVE SCIENCES, 2003, 7 (11) : 498 - 504
  • [8] Predicting Cognitive State from Eye Movements
    Henderson, John M.
    Shinkareva, Svetlana V.
    Wang, Jing
    Luke, Steven G.
    Olejarczyk, Jenn
    [J]. PLOS ONE, 2013, 8 (05):
  • [9] Towards Reading Trackers in the Wild: Detecting Reading Activities by EOG Glasses and Deep Neural Networks
    Ishimaru, Shoya
    Kise, Koichi
    Hoshika, Kensuke
    Dengel, Andreas
    Kunze, Kai
    [J]. PROCEEDINGS OF THE 2017 ACM INTERNATIONAL JOINT CONFERENCE ON PERVASIVE AND UBIQUITOUS COMPUTING AND PROCEEDINGS OF THE 2017 ACM INTERNATIONAL SYMPOSIUM ON WEARABLE COMPUTERS (UBICOMP/ISWC '17 ADJUNCT), 2017, : 704 - 711
  • [10] Kelton Conor, 2019, 2019 S EYE TRACK RES