A Human-Robot Interaction System Calculating Visual Focus of Human's Attention Level

被引:19
作者
Chakraborty, Partha [1 ]
Ahmed, Sabbir [2 ]
Abu Yousuf, Mohammad [2 ]
Azad, Akm [3 ]
Alyami, Salem A. [4 ]
Moni, Mohammad Ali [5 ]
机构
[1] Comilla Univ, Dept Comp Sci & Engn, Cumilla 3506, Bangladesh
[2] Jahangimagar Univ, Inst Informat Technol, Dhaka 1342, Bangladesh
[3] Univ New South Wales Sydney UNSW Sydney, Sch Biotechnol & Biomol Sci, Sydney, NSW 2052, Australia
[4] Imam Mohammad Ibn Saud Islamic Univ IMSIU, Fac Sci, Dept Math & Stat, Riyadh 13318, Saudi Arabia
[5] Univ New South Wales Sydney UNSW Sydney, WHO Collaborating Ctr eHlth, UNSW Digital Hlth, Sydney, NSW 2052, Australia
关键词
Head; Visualization; Tracking; Writing; Feature extraction; Task analysis; Lighting; Human-robot interaction; attention level; visual focus of attention; concentration; ANN; RNN-LSTM; EYE-TRACKING; SENTIMENT;
D O I
10.1109/ACCESS.2021.3091642
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Attention is the mental awareness of human on a particular object or a piece of information. The level of attention indicates how intense the focus is on an object or an instance. In this study, several types of human attention level have been observed. After introducing image segmentation and detection technique for facial features, eyeball movement and gaze estimation were measured. Eye movement were assessed using the video data, and a total of 10197 data instances were manually labelled for the attention level. Then Artificial Neural Network (ANN) and Recurrent Neural Network-Long Short Term Memory (LSTM) based Deep learning (DL) architectures have been proposed for analysing the data. Next, the trained DL model has been implanted into a robotic system that is capable of detecting various features; ultimately leading to the calculation of visual attention for reading, browsing, and writing purposes. This system is capable of checking the attention level of the participants and also can detect if participants are present or not. Based on a certain level of visual focus of attention (VFOA), this system interacts with the person, generates awareness and establishes verbal or visual communication with that person. The proposed ML techniques have achieved almost 99.24% validation accuracy and 99.43% test accuracy. It is also shown in the comparative study that, since the dataset volumes are limited, ANN is more suitable for attention level calculation than RNN-LSTM. We hope that the implemented robotic structure manifests the real-world implication of the proposed method.
引用
收藏
页码:93409 / 93421
页数:13
相关论文
共 45 条
[1]  
Abbas SSA, 2017, 2017 2ND IEEE INTERNATIONAL CONFERENCE ON WIRELESS COMMUNICATIONS, SIGNAL PROCESSING AND NETWORKING (WISPNET), P1160, DOI 10.1109/WiSPNET.2017.8299946
[2]   Designing a low-cost eyeball tracking keyboard for paralyzed people [J].
Al-Kassim, Zeenat ;
Memon, Qurban Ali .
COMPUTERS & ELECTRICAL ENGINEERING, 2017, 58 :20-29
[3]   Implementation of a Character Recognition System Based on Finger-Joint Tracking Using a Depth Camera [J].
Alam, Md. Shahinur ;
Kwon, Ki-Chul ;
Kim, Nam .
IEEE TRANSACTIONS ON HUMAN-MACHINE SYSTEMS, 2021, 51 (03) :229-241
[4]   Trajectory-Based Air-Writing Recognition Using Deep Neural Network and Depth Sensor [J].
Alam, Md. Shahinur ;
Kwon, Ki-Chul ;
Alam, Md. Ashraful ;
Abbass, Mohammed Y. ;
Imtiaz, Shariar Md ;
Kim, Nam .
SENSORS, 2020, 20 (02)
[5]   Adam Deep Learning With SOM for Human Sentiment Classification [J].
Ali, Md Nawab Yousuf ;
Sarowar, Md Golam ;
Rahman, Md Lizur ;
Chaki, Jyotismita ;
Dey, Nilanjan ;
Tavares, Joao Manuel R. S. .
INTERNATIONAL JOURNAL OF AMBIENT COMPUTING AND INTELLIGENCE, 2019, 10 (03) :92-116
[6]  
[Anonymous], 2014, P WORKSH MULT MULT P
[7]  
[Anonymous], 2011, P INT S INT SIGN PRO
[8]   Blink detection and eye tracking for eye localization [J].
Bhaskar, TN ;
Keat, FT ;
Ranganath, S ;
Venkatesh, YV .
IEEE TENCON 2003: CONFERENCE ON CONVERGENT TECHNOLOGIES FOR THE ASIA-PACIFIC REGION, VOLS 1-4, 2003, :821-824
[9]  
Bhoir SA., 2015, P 11 CONSTR SPEC C, P222
[10]  
Canedo D., 2018, P HIGHL PRACT APPL A, P371