Multi-modal fusion learning through biosignal, audio, and visual content for detection of mental stress

被引:6
|
作者
Dogan, Gulin [1 ]
Akbulut, Fatma Patlar [2 ]
机构
[1] Istanbul Kultur Univ, Dept Comp Engn, TR-34158 Istanbul, Turkiye
[2] Istanbul Kultur Univ, Dept Software Engn, TR-34158 Istanbul, Turkiye
来源
NEURAL COMPUTING & APPLICATIONS | 2023年 / 35卷 / 34期
关键词
Stress detection; Sequential and non-sequential model; Fine-tuning; Multi-modality; MOMENTARY ASSESSMENT; RECOGNITION; VOICE; FACE;
D O I
10.1007/s00521-023-09036-4
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Mental stress is a significant risk factor for several maladies and can negatively impact a person's quality of life, including their work and personal relationships. Traditional methods of detecting mental stress through interviews and questionnaires may not capture individuals' instantaneous emotional responses. In this study, the method of experience sampling was used to analyze the participants' immediate affective responses, which provides a more comprehensive and dynamic understanding of the participants' experiences. WorkStress3D dataset was compiled using information gathered from 20 participants for three distinct modalities. During an average of one week, 175 h of data containing physiological signals such as BVP, EDA, and body temperature, as well as facial expressions and auditory data, were collected from a single subject. We present a novel fusion model that uses double-early fusion approaches to combine data from multiple modalities. The model's F1 score of 0.94 with a loss of 0.18 is very encouraging, showing that it can accurately identify and classify varying degrees of stress. Furthermore, we investigate the utilization of transfer learning techniques to improve the efficacy of our stress detection system. Despite our efforts, we were unable to attain better results than the fusion model. Transfer learning resulted in an accuracy of 0.93 and a loss of 0.17, illustrating the difficulty of adapting pre-trained models to the task of stress analysis. The results we obtained emphasize the significance of multi-modal fusion in stress detection and the importance of selecting the most suitable model architecture for the given task. The proposed fusion model demonstrates its potential for achieving an accurate and robust classification of stress. This research contributes to the field of stress analysis and contributes to the development of effective models for stress detection.
引用
收藏
页码:24435 / 24454
页数:20
相关论文
共 50 条
  • [21] Multi-modal fusion in ergonomic health: bridging visual and pressure for sitting posture detection
    Quan, Qinxiao
    Gao, Yang
    Bai, Yang
    Jin, Zhanpeng
    CCF TRANSACTIONS ON PERVASIVE COMPUTING AND INTERACTION, 2024, : 380 - 393
  • [22] Multi-modal audio-visual event recognition for football analysis
    Barnard, M
    Odobez, JM
    Bengio, S
    2003 IEEE XIII WORKSHOP ON NEURAL NETWORKS FOR SIGNAL PROCESSING - NNSP'03, 2003, : 469 - 478
  • [23] Multi-Modal fusion with multi-level attention for Visual Dialog
    Zhang, Jingping
    Wang, Qiang
    Han, Yahong
    INFORMATION PROCESSING & MANAGEMENT, 2020, 57 (04)
  • [24] Multi-modal authentication system based on audio-visual data
    Debnath, Saswati
    Roy, Pinki
    PROCEEDINGS OF THE 2019 IEEE REGION 10 CONFERENCE (TENCON 2019): TECHNOLOGY, KNOWLEDGE, AND SOCIETY, 2019, : 2507 - 2512
  • [25] MULTI-MODAL FUSION LEARNING FOR CERVICAL DYSPLASIA DIAGNOSIS
    Chen, Tingting
    Ma, Xinjun
    Ying, Xingde
    Wang, Wenzhe
    Yuan, Chunnv
    Lu, Weiguo
    Chen, Danny Z.
    Wu, Jian
    2019 IEEE 16TH INTERNATIONAL SYMPOSIUM ON BIOMEDICAL IMAGING (ISBI 2019), 2019, : 1505 - 1509
  • [26] Memory based fusion for multi-modal deep learning
    Priyasad, Darshana
    Fernando, Tharindu
    Denman, Simon
    Sridharan, Sridha
    Fookes, Clinton
    INFORMATION FUSION, 2021, 67 : 136 - 146
  • [27] Multi-modal Graph and Sequence Fusion Learning for Recommendation
    Wang, Zejun
    Wu, Xinglong
    Yang, Hongwei
    He, Hui
    Tai, Yu
    Zhang, Weizhe
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2023, PT I, 2024, 14425 : 357 - 369
  • [28] MMFusion: A Generalized Multi-Modal Fusion Detection Framework
    Cui, Leichao
    Li, Xiuxian
    Meng, Min
    Mo, Xiaoyu
    2023 IEEE INTERNATIONAL CONFERENCE ON DEVELOPMENT AND LEARNING, ICDL, 2023, : 415 - 422
  • [29] Improving multi-modal data fusion by anomaly detection
    Jakub Simanek
    Vladimir Kubelka
    Michal Reinstein
    Autonomous Robots, 2015, 39 : 139 - 154
  • [30] Improving multi-modal data fusion by anomaly detection
    Simanek, Jakub
    Kubelka, Vladimir
    Reinstein, Michal
    AUTONOMOUS ROBOTS, 2015, 39 (02) : 139 - 154