Expression Recognition Survey Through Multi-Modal Data Analytics

被引:0
作者
Ramyasree, Kummari [1 ,2 ]
Kumar, Ch. Sumanth [1 ]
机构
[1] GITAM Deemed Univ, Dept E&ECE, Visakhapatnam, AP, India
[2] Guru Nanak Inst Tech Campus, Hyderabad, TS, India
来源
INTERNATIONAL JOURNAL OF COMPUTER SCIENCE AND NETWORK SECURITY | 2022年 / 22卷 / 06期
关键词
Expression recognition; speech; face; MFCC; Prosodic; LBP; Machine Learning; Databases; AUDIOVISUAL EMOTION RECOGNITION; FACIAL-EXPRESSION; SPEECH;
D O I
10.22937/IJCSNS.2022.22.6.74
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In computer vision, capturing human expression from a video is essential for various time-sensitive applications, including driver safety and education. Because there are multiple expression models (e.g., speech, face, gesture, etc. ), Human Expression Recognition can be done in various methods. In this work, we look at multiple strategies for recognizing facial expressions. We primarily concentrated on two models in this review: speech signal and facial image. The entrance survey is conducted in two steps, according to the generic approach of human expression recognition.The methods we used to extract features were classified into two groups: audio features and facial features. The review found different classifiers helped to decide the best framing of the Markov chain for a model. We also looked at specific multimodal expression recognition algorithms that used the multimodal notion at various stages, such as feature fusion and decision fusion. We also looked at several databases that have been used by previous studies in addition to these approaches. A full-fledged comparison is also provided to show the review ultimately.
引用
收藏
页码:600 / 610
页数:11
相关论文
共 106 条
[1]   Face description with local binary patterns:: Application to face recognition [J].
Ahonen, Timo ;
Hadid, Abdenour ;
Pietikainen, Matti .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2006, 28 (12) :2037-2041
[2]   Spoken emotion recognition using hierarchical classifiers [J].
Albornoz, Enrique M. ;
Milone, Diego H. ;
Rufiner, Hugo L. .
COMPUTER SPEECH AND LANGUAGE, 2011, 25 (03) :556-570
[3]  
Anina I., 2015, IEEE INT CONF AUTOMA
[4]  
[Anonymous], 2006, 22 INT C DAT ENG WOR
[5]  
[Anonymous], 2013, P 21 ACM INT C MULTI
[6]  
[Anonymous], 2014, REAL TIME EMOTION RE
[7]  
Antonini G, 2006, LECT NOTES COMPUT SC, V4179, P710
[8]   Multimodal Intelligent Eye-Gaze Tracking System [J].
Biswas, Pradipta ;
Langdon, Pat .
INTERNATIONAL JOURNAL OF HUMAN-COMPUTER INTERACTION, 2015, 31 (04) :277-294
[9]   Accuracy vs. Availability Heuristic in Multimodal Affect Detection in the Wild [J].
Bosch, Nigel ;
Chen, Huili ;
Baker, Ryan ;
Shute, Valerie ;
D'Mello, Sidney .
ICMI'15: PROCEEDINGS OF THE 2015 ACM INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION, 2015, :267-274
[10]   IEMOCAP: interactive emotional dyadic motion capture database [J].
Busso, Carlos ;
Bulut, Murtaza ;
Lee, Chi-Chun ;
Kazemzadeh, Abe ;
Mower, Emily ;
Kim, Samuel ;
Chang, Jeannette N. ;
Lee, Sungbok ;
Narayanan, Shrikanth S. .
LANGUAGE RESOURCES AND EVALUATION, 2008, 42 (04) :335-359