Deep learning analysis of mobile physiological, environmental and location sensor data for emotion detection

被引:160
作者
Kanjo, Eiman [1 ]
Younis, Eman M. G. [2 ]
Ang, Chee Siang [3 ]
机构
[1] Nottingham Trent Univ, Comp & Technol Dept, Nottingham, England
[2] Minia Univ, Fac Comp & Informat, Al Minya, Egypt
[3] Univ Kent, Sch Engn & Digital Arts, Canterbury, Kent, England
关键词
Deep learning; Emotion recognition; Convoltutional neural network; Long short-term memory mobile sensing; RECOGNITION;
D O I
10.1016/j.inffus.2018.09.001
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The detection and monitoring of emotions are important in various applications, e.g., to enable naturalistic and personalised human-robot interaction. Emotion detection often require modelling of various data inputs from multiple modalities, including physiological signals (e.g., EEG and GSR), environmental data (e.g., audio and weather), videos (e.g., for capturing facial expressions and gestures) and more recently motion and location data. Many traditional machine learning algorithms have been utilised to capture the diversity of multimodal data at the sensors and features levels for human emotion classification. While the feature engineering processes often embedded in these algorithms are beneficial for emotion modelling, they inherit some critical limitations which may hinder the development of reliable and accurate models. In this work, we adopt a deep learning approach for emotion classification through an iterative process by adding and removing large number of sensor signals from different modalities. Our dataset was collected in a real-world study from smart-phones and wearable devices. It merges local interaction of three sensor modalities: on-body, environmental and location into global model that represents signal dynamics along with the temporal relationships of each modality. Our approach employs a series of learning algorithms including a hybrid approach using Convolutional Neural Network and Long Short-term Memory Recurrent Neural Network (CNN-LSTM) on the raw sensor data, eliminating the needs for manual feature extraction and engineering. The results show that the adoption of deep-learning approaches is effective in human emotion classification when large number of sensors input is utilised (average accuracy 95% and F-Measure=%95) and the hybrid models outperform traditional fully connected deep neural network (average accuracy 73% and F-Measure=73%). Furthermore, the hybrid models outperform previously developed Ensemble algorithms that utilise feature engineering to train the model average accuracy 83% and F-Measure=82%)
引用
收藏
页码:46 / 56
页数:11
相关论文
共 46 条
  • [1] ECG Pattern Analysis for Emotion Detection
    Agrafioti, Foteini
    Hatzinakos, Dimitrios
    Anderson, Adam K.
    [J]. IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2012, 3 (01) : 102 - 115
  • [2] NeuroPlace: Categorizing urban places according to mental states
    Al-barrak, Lulwah
    Kanjo, Eiman
    Younis, Eman M. G.
    [J]. PLOS ONE, 2017, 12 (09):
  • [3] Shopmobia: An Emotion-based Shop Rating System
    Alajmi, Nouf
    Kanjo, Eiman
    El Mawass, Nour
    Chamberlain, Alan
    [J]. 2013 HUMAINE ASSOCIATION CONFERENCE ON AFFECTIVE COMPUTING AND INTELLIGENT INTERACTION (ACII), 2013, : 745 - 750
  • [4] [Anonymous], P 25 INT JOINT C ART, DOI DOI 10.48550/ARXIV.1604.08880
  • [5] [Anonymous], 2010, Advances in neural information processing systems
  • [6] [Anonymous], P 3 INT C LEARNING R
  • [7] [Anonymous], P INT C MULT INT
  • [8] [Anonymous], 2016, ARXIV160603238
  • [9] [Anonymous], P AAAI WORKSH ART IN
  • [10] [Anonymous], PERS UBIQUITOUS COMP