A Vision-Based System for Monitoring Elderly People at Home

被引:47
作者
Buzzelli, Marco [1 ]
Albe, Alessio [1 ]
Ciocca, Gianluigi [1 ]
机构
[1] Univ Milano Bicocca, Dept Comp Sci Syst & Commun, Viale Sarca 336, I-20126 Milan, Italy
来源
APPLIED SCIENCES-BASEL | 2020年 / 10卷 / 01期
关键词
computer vision; action recognition; deep learning; internet of things; assisted living; ACTION RECOGNITION; CARE;
D O I
10.3390/app10010374
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
Assisted living technologies can be of great importance for taking care of elderly people and helping them to live independently. In this work, we propose a monitoring system designed to be as unobtrusive as possible, by exploiting computer vision techniques and visual sensors such as RGB cameras. We perform a thorough analysis of existing video datasets for action recognition, and show that no single dataset can be considered adequate in terms of classes or cardinality. We subsequently curate a taxonomy of human actions, derived from different sources in the literature, and provide the scientific community with considerations about the mutual exclusivity and commonalities of said actions. This leads us to collecting and publishing an aggregated dataset, called ALMOND (Assisted Living MONitoring Dataset), which we use as the training set for a vision-based monitoring approach.We rigorously evaluate our solution in terms of recognition accuracy using different state-of-the-art architectures, eventually reaching 97% on inference of basic poses, 83% on alerting situations, and 71% on daily life actions. We also provide a general methodology to estimate the maximum allowed distance between camera and monitored subject. Finally, we integrate the defined actions and the trained model into a computer-vision-based application, specifically designed for the objective of monitoring elderly people at their homes.
引用
收藏
页数:25
相关论文
共 62 条
[41]  
Ng JYH, 2015, PROC CVPR IEEE, P4694, DOI 10.1109/CVPR.2015.7299101
[42]   Learning Spatio-Temporal Representation with Pseudo-3D Residual Networks [J].
Qiu, Zhaofan ;
Yao, Ting ;
Mei, Tao .
2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, :5534-5542
[43]   Histogram of Oriented Principal Components for Cross-View Action Recognition [J].
Rahmani, Hossein ;
Mahmood, Arif ;
Du Huynh ;
Mian, Ajmal .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2016, 38 (12) :2430-2443
[44]   Recognizing 50 human action categories of web videos [J].
Reddy, Kishore K. ;
Shah, Mubarak .
MACHINE VISION AND APPLICATIONS, 2013, 24 (05) :971-981
[45]  
Roser M, 2019, Human height. Our world in data
[46]   NTU RGB plus D: A Large Scale Dataset for 3D Human Activity Analysis [J].
Shahroudy, Amir ;
Liu, Jun ;
Ng, Tian-Tsong ;
Wang, Gang .
2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, :1010-1019
[47]  
Sigurdsson G.A., 2016, P EUR C COMP VIS AMS, P20
[48]  
Simonyan Karen, 2014, ADV NEURAL INFORM PR, P2, DOI [DOI 10.1002/14651858.CD001941.PUB3, DOI 10.1002/14651858.00001941.PUB3]
[49]   Spatio-Temporal Attention-Based LSTM Networks for 3D Action Recognition and Detection [J].
Song, Sijie ;
Lan, Cuiling ;
Xing, Junliang ;
Zeng, Wenjun ;
Liu, Jiaying .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2018, 27 (07) :3459-3471
[50]  
Soomro K., 2012, UCF101, V2, P1