A Hybrid Multimodal Emotion Recognition Framework for UX Evaluation Using Generalized Mixture Functions

被引:7
作者
Razzaq, Muhammad Asif [1 ,2 ]
Hussain, Jamil [3 ]
Bang, Jaehun [4 ]
Hua, Cam-Hao [2 ]
Satti, Fahad Ahmed [2 ,5 ]
Rehman, Ubaid Ur [2 ,5 ]
Bilal, Hafiz Syed Muhammad [5 ]
Kim, Seong Tae [2 ]
Lee, Sungyoung [2 ]
机构
[1] Fatima Jinnah Women Univ, Dept Comp Sci, Rawalpindi 46000, Pakistan
[2] Kyung Hee Univ, Dept Comp Sci & Engn, Ubiquitous Comp Lab, Yongin 17104, South Korea
[3] Sejong Univ, Dept Data Sci, Seoul 30019, South Korea
[4] Hanwha Corp Momentum, Hanwha Bldg, 86 Cheonggyecheon Ro, Seoul 04541, South Korea
[5] Natl Univ Sci & Technol NUST, Sch Elect Engn & Comp Sci SEECS, Dept Comp, Islamabad 44000, Pakistan
关键词
emotion recognition; user experience; audio-based emotion recognition; feature fusioning; decision fusioning; generalized mixture function; FUSION; NETWORK; MODEL;
D O I
10.3390/s23094373
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
Multimodal emotion recognition has gained much traction in the field of affective computing, human-computer interaction (HCI), artificial intelligence (AI), and user experience (UX). There is growing demand to automate analysis of user emotion towards HCI, AI, and UX evaluation applications for providing affective services. Emotions are increasingly being used, obtained through the videos, audio, text or physiological signals. This has led to process emotions from multiple modalities, usually combined through ensemble-based systems with static weights. Due to numerous limitations like missing modality data, inter-class variations, and intra-class similarities, an effective weighting scheme is thus required to improve the aforementioned discrimination between modalities. This article takes into account the importance of difference between multiple modalities and assigns dynamic weights to them by adapting a more efficient combination process with the application of generalized mixture (GM) functions. Therefore, we present a hybrid multimodal emotion recognition (H-MMER) framework using multi-view learning approach for unimodal emotion recognition and introducing multimodal feature fusion level, and decision level fusion using GM functions. In an experimental study, we evaluated the ability of our proposed framework to model a set of four different emotional states (Happiness, Neutral, Sadness, and Anger) and found that most of them can be modeled well with significantly high accuracy using GM functions. The experiment shows that the proposed framework can model emotional states with an average accuracy of 98.19% and indicates significant gain in terms of performance in contrast to traditional approaches. The overall evaluation results indicate that we can identify emotional states with high accuracy and increase the robustness of an emotion classification system required for UX measurement.
引用
收藏
页数:25
相关论文
共 54 条
[1]   Multimodal spatiotemporal skeletal kinematic gait feature fusion for vision-based fall detection [J].
Amsaprabhaa, M. ;
Jane, Y. Nancy ;
Nehemiah, H. Khanna .
EXPERT SYSTEMS WITH APPLICATIONS, 2023, 212
[2]   Multimodal Machine Learning: A Survey and Taxonomy [J].
Baltrusaitis, Tadas ;
Ahuja, Chaitanya ;
Morency, Louis-Philippe .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2019, 41 (02) :423-443
[3]   Adaptive Data Boosting Technique for Robust Personalized Speech Emotion in Emotionally-Imbalanced Small-Sample Environments [J].
Bang, Jaehun ;
Hur, Taeho ;
Kim, Dohyeong ;
Huynh-The, Thien ;
Lee, Jongwon ;
Han, Yongkoo ;
Banos, Oresti ;
Kim, Jee-In ;
Lee, Sungyoung .
SENSORS, 2018, 18 (11)
[4]  
Beliakov G, 2016, STUD FUZZ SOFT COMP, V329, P1, DOI 10.1007/978-3-319-24753-3
[5]   Deep learning-based natural language sentiment classification model for recognizing users' sentiments toward residential space [J].
Chang, Sun-Woo ;
Dong, Won-Hyeok ;
Rhee, Deuk-Young ;
Jun, Han-Jong .
ARCHITECTURAL SCIENCE REVIEW, 2021, 64 (05) :410-421
[6]   Cross-Subject Multimodal Emotion Recognition Based on Hybrid Fusion [J].
Cimtay, Yucel ;
Ekmekcioglu, Erhan ;
Caglar-Ozhan, Seyma .
IEEE ACCESS, 2020, 8 :168865-168878
[7]   Combining multiple algorithms in classifier ensembles using generalized mixture functions [J].
Costa, Valdigleis S. ;
Farias, Antonio Diego S. ;
Bedregal, Benjamin ;
Santiago, Regivan H. N. ;
Canuto, Anne Magaly de P. .
NEUROCOMPUTING, 2018, 313 :402-414
[8]   Multiscale Amplitude Feature and Significance of Enhanced Vocal Tract Information for Emotion Classification [J].
Deb, Suman ;
Dandapat, Samarendra .
IEEE TRANSACTIONS ON CYBERNETICS, 2019, 49 (03) :802-815
[9]  
deeplearning4j, DEEP LEARN LIB JAV
[10]   A Novel Emotion-Aware Method Based on the Fusion of Textual Description of Speech, Body Movements, and Facial Expressions [J].
Du, Guanglong ;
Zeng, Yuwen ;
Su, Kang ;
Li, Chunquan ;
Wang, Xueqian ;
Teng, Shaohua ;
Li, Di ;
Liu, Peter Xiaoping .
IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2022, 71