Detecting Screen Presence with Activity-Oriented RGB Camera in Egocentric Videos

被引:0
作者
Adate, Amit [1 ]
Shahi, Soroush [2 ,3 ]
Alharbi, Rawan [2 ,3 ]
Sen, Sougata [4 ]
Gao, Yang [2 ,3 ]
Katsaggelos, Aggelos K. [1 ,2 ]
Alshurafa, Nabil [2 ,3 ]
机构
[1] Northwestern Univ, Dept Elect & Comp Engn, Evanston, IL 60208 USA
[2] Northwestern Univ, Dept Comp Sci, Evanston, IL USA
[3] Northwestern Univ, Dept Prevent Med, Chicago, IL 60611 USA
[4] BITS, Dept Comp Sci & Informat Syst, Pilani, Goa, India
来源
2022 IEEE INTERNATIONAL CONFERENCE ON PERVASIVE COMPUTING AND COMMUNICATIONS WORKSHOPS AND OTHER AFFILIATED EVENTS (PERCOM WORKSHOPS) | 2022年
基金
美国国家科学基金会;
关键词
Object Detection; Egocentric Videos; Fisheye Lens; Wearable Camera;
D O I
10.1109/PerComWorkshops53856.2022.9767433
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Screen time is associated with several health risk behaviors including mindless eating, sedentary behavior, and decreased academic performance. Screen time behavior is traditionally assessed with self-report measures, which are known to be burdensome, inaccurate, and imprecise. Recent methods to automatically detect screen time are geared more towards detecting television screens from wearable cameras that record high-resolution video. Activity-oriented wearable cameras (i.e., cameras oriented towards the wearer with a fisheye lens) have recently been designed and shown to reduce privacy concerns, yet pose a greater challenge in capturing screens due to their orientation and fewer pixels on target. Methods that detect screens from low-power, low-resolution wearable camera video are needed given the increased adoption of such devices in longitudinal studies. We propose a method that leverages deep learning algorithms and lower-resolution images from an activity-oriented camera to detect screen presence from multiple types of screens with high variability of pixel on target (e.g., near and far TV, smartphones, laptops, and tablets). We test our system in a real-world study comprising 10 individuals, 80 hours of data, and 1.2 million low-resolution RGB frames. Our results outperform existing state-of-the-art video screen detection methods yielding an F1-score of 81%. This paper demonstrates the potential for detecting screen-watching behavior in longitudinal studies using activity-oriented cameras, paving the way for a nuanced understanding of screen time's relationship with health risk behaviors.
引用
收藏
页数:6
相关论文
共 20 条
[1]  
[Anonymous], 2021, COCO MINIVAL BENCHMA
[2]  
[Anonymous], 2020, PANASONICGRIDEYE
[3]  
[Anonymous], 2021, OV2640 OMNIVISION
[4]  
Apaolaza A., 2016, P 2016 CHI C HUM FAC, P2766
[5]  
Apaolaza A., 2014, UNDERSTANDING DIVISI
[6]   Mindfulness, Compulsive Mobile Social Media Use, and Derived Stress: The Mediating Roles of Self-Esteem and Social Anxiety [J].
Apaolaza, Vanessa ;
Hartmann, Patrick ;
D'Souza, Clare ;
Gilsanz, Ainhize .
CYBERPSYCHOLOGY BEHAVIOR AND SOCIAL NETWORKING, 2019, 22 (06) :388-396
[7]   Egocentric Gesture Recognition Using Recurrent 3D Convolutional Neural Networks with Spatiotemporal Transformer Modules [J].
Cao, Congqi ;
Zhang, Yifan ;
Wu, Yi ;
Lu, Hanqing ;
Cheng, Jian .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, :3783-3791
[8]  
Dingsheng Deng, 2020, 2020 7th International Forum on Electrical Engineering and Automation (IFEEA), P949, DOI 10.1109/IFEEA51475.2020.00199
[9]   Enhancing Lifelogging Privacy by Detecting Screens [J].
Korayem, Mohammed ;
Templeman, Robert ;
Chen, Dennis ;
Crandall, David ;
Kapadia, Apu .
34TH ANNUAL CHI CONFERENCE ON HUMAN FACTORS IN COMPUTING SYSTEMS, CHI 2016, 2016, :4309-4314
[10]  
Li Z., 2018, DetNet: A Backbone network for Object Detection