A Vision-Based System for Monitoring Elderly People at Home

被引:47
作者
Buzzelli, Marco [1 ]
Albe, Alessio [1 ]
Ciocca, Gianluigi [1 ]
机构
[1] Univ Milano Bicocca, Dept Comp Sci Syst & Commun, Viale Sarca 336, I-20126 Milan, Italy
来源
APPLIED SCIENCES-BASEL | 2020年 / 10卷 / 01期
关键词
computer vision; action recognition; deep learning; internet of things; assisted living; ACTION RECOGNITION; CARE;
D O I
10.3390/app10010374
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
Assisted living technologies can be of great importance for taking care of elderly people and helping them to live independently. In this work, we propose a monitoring system designed to be as unobtrusive as possible, by exploiting computer vision techniques and visual sensors such as RGB cameras. We perform a thorough analysis of existing video datasets for action recognition, and show that no single dataset can be considered adequate in terms of classes or cardinality. We subsequently curate a taxonomy of human actions, derived from different sources in the literature, and provide the scientific community with considerations about the mutual exclusivity and commonalities of said actions. This leads us to collecting and publishing an aggregated dataset, called ALMOND (Assisted Living MONitoring Dataset), which we use as the training set for a vision-based monitoring approach.We rigorously evaluate our solution in terms of recognition accuracy using different state-of-the-art architectures, eventually reaching 97% on inference of basic poses, 83% on alerting situations, and 71% on daily life actions. We also provide a general methodology to estimate the maximum allowed distance between camera and monitored subject. Finally, we integrate the defined actions and the trained model into a computer-vision-based application, specifically designed for the objective of monitoring elderly people at their homes.
引用
收藏
页数:25
相关论文
共 62 条
[11]  
Bourouis Abderrahim, 2011, International Journal of Computer Science & Information Technology, V3, P74, DOI 10.5121/ijcsit.2011.3306
[12]   Quo Vadis, Action Recognition? A New Model and the Kinetics Dataset [J].
Carreira, Joao ;
Zisserman, Andrew .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :4724-4733
[13]   Dynamic key-frame extraction for video summarization [J].
Ciocca, G ;
Schettini, R .
INTERNET IMAGING VI, 2005, 5670 :137-142
[14]  
Ciocca G., 2006, J REAL-TIME IMAGE PR, V1, P69, DOI DOI 10.1007/s11554-012-0278-1
[15]   Elder Tracking and Fall Detection System Using Smart Tiles [J].
Daher, Mohamad ;
Diab, Ahmad ;
El Najjar, Maan El Badaoui ;
Khalil, Mohamad Ali ;
Charpillet, Francois .
IEEE SENSORS JOURNAL, 2017, 17 (02) :469-479
[16]  
Diba A., 2017, Temporal 3d convnets: New architecture and transfer learning for video classification
[17]  
Donahue J, 2015, PROC CVPR IEEE, P2625, DOI 10.1109/CVPR.2015.7298878
[18]   Learning Spatiotemporal Features with 3D Convolutional Networks [J].
Du Tran ;
Bourdev, Lubomir ;
Fergus, Rob ;
Torresani, Lorenzo ;
Paluri, Manohar .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, :4489-4497
[19]  
European Commission - Economic and Financial Affairs, 2018, 2018 AG REP
[20]   Convolutional Two-Stream Network Fusion for Video Action Recognition [J].
Feichtenhofer, Christoph ;
Pinz, Axel ;
Zisserman, Andrew .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :1933-1941