Modelling an efficient hybridized approach for facial emotion recognition using unconstraint videos and deep learning approaches

被引:0
作者
Bhushanam, P. Naga [1 ]
Kumar, S. Selva [1 ]
机构
[1] VIT AP Univ, Sch Comp Sci & Engn, Amaravati, Andhra Pradesh, India
关键词
Facial emotion recognition; Deep learning; Unconstraint video; Multi-modal; Prediction accuracy;
D O I
10.1007/s00500-024-09668-1
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Facial Emotion detection (FER) is primarily used to assess human emotion to meet the demands of several real-time applications, including emotion detection, computer-human interfaces, biometrics, forensics, and human-robot collaboration. However, several current techniques fall short of providing accurate predictions with a low error rate. This study focuses on modeling an effective FER with unrestricted videos using a hybrid SegNet and ConvNet model. SegNet is used to segment the regions of facial expression, and ConvNet is used to analyze facial features and to make predictions about emotions like surprise, sadness, and happiness, among others. The suggested hybridized approach uses a neural network model to classify face characteristics depending on their location. The proposed model aims to recognize facial emotions with a quicker convergence rate and improved prediction accuracy. This work takes into account the internet-accessible datasets from the FER2013, Kaggle, and GitHub databases to execute execution. To accomplish generalization and improve the quality of prediction, the model acts as a multi-modal application. With the available datasets throughout the testing procedure, the suggested model provides 95% prediction accuracy. Additionally, the suggested hybridized model is used to calculate the system's importance. The experimental results show that, in comparison to previous techniques, the expected model provides superior prediction results and produces better trade-offs. Other related statistical measures are also assessed and contrasted while the simulation is being run in the MATLAB 2020a environment.
引用
收藏
页码:3823 / 3846
页数:24
相关论文
共 29 条
[1]   Ensemble of CNN for multi-focus image fusion [J].
Amin-Naji, Mostafa ;
Aghagolzadeh, Ali ;
Ezoji, Mehdi .
INFORMATION FUSION, 2019, 51 :201-214
[2]   Facial Expression Recognition Using Computer Vision: A Systematic Review [J].
Canedo, Daniel ;
Neves, Antonio J. R. .
APPLIED SCIENCES-BASEL, 2019, 9 (21)
[3]   Xception: Deep Learning with Depthwise Separable Convolutions [J].
Chollet, Francois .
30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, :1800-1807
[4]   EmotiW 2016: Video and Group-Level Emotion Recognition Challenges [J].
Dhall, Abhinav ;
Goecke, Roland ;
Joshi, Jyoti ;
Hoey, Jesse ;
Gedeon, Tom .
ICMI'16: PROCEEDINGS OF THE 18TH ACM INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION, 2016, :427-432
[5]   Facial Expression Recognition Using Convolutional Neural Network [J].
Gan, Yijun .
PROCEEDINGS OF THE 2ND INTERNATIONAL CONFERENCE ON VISION, IMAGE AND SIGNAL PROCESSING (ICVISP 2018), 2018,
[6]  
Hans A.S.A., 2021, Int. J. Adv. Signal Image Sci, V7, P11, DOI DOI 10.29284/IJASIS.7.1.2021.11-20
[7]  
He K., 2016, PROC CVPR IEEE, DOI [DOI 10.1109/CVPR.2016.90, 10.1109/CVPR.2016.90]
[8]  
Hussain SA, 2020, J. Phys. Conf. Ser, V1432, DOI [10.1088/1742-6596/1432/1/012087, DOI 10.1088/1742-6596/1432/1/012087]
[9]   Caffe: Convolutional Architecture for Fast Feature Embedding [J].
Jia, Yangqing ;
Shelhamer, Evan ;
Donahue, Jeff ;
Karayev, Sergey ;
Long, Jonathan ;
Girshick, Ross ;
Guadarrama, Sergio ;
Darrell, Trevor .
PROCEEDINGS OF THE 2014 ACM CONFERENCE ON MULTIMEDIA (MM'14), 2014, :675-678
[10]  
Jiang JL, 2014, THUMOS CHALLENGE ACT