NA-Resnet: neighbor block and optimized attention module for global-local feature extraction in facial expression recognition

被引:2
作者
Qi, Yongfeng [1 ]
Zhou, Chenyang [1 ]
Chen, Yixing [2 ]
机构
[1] Northwest Normal Univ, Coll Comp Sci & Engn, Lanzhou 730070, Gansu, Peoples R China
[2] Guangzhou Univ, Sch Comp Sci & Cyber Engn, Guangzhou 510006, Guangdong, Peoples R China
基金
中国国家自然科学基金;
关键词
Deep learning; Facial expression recognition; Convolutional neural network (CNN); NA-Resnet; Neighbor block; Optimized attention module; REPRESENTATION;
D O I
10.1007/s11042-022-14191-2
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
As deep networks constantly deepen to extract high-level abstract features, the significance of shallow features for the target task will inevitably diminish. To address this issue and provide novel technical support for current research in the field of facial expression recognition (FER), in this article, we propose a network that can increase the decision weight of the shallow and middle feature mappings through the neighbor block (Nei Block) and concentrate on the crucial areas for extracting necessary features through the optimized attention module (OAM), called NA-Resnet. Our work has several merits. First, to the best of our knowledge, NA-Resnet is the first network that directly utilizes surface features to assist image classification. Second, the suggested OAM is embedded into each layer of the network that can precisely extract critical information appropriate to the current stage. Third, our model achieves the best exhibition when using a single relatively lightweight network without a network ensemble on Fer2013. Extensive experiments have been conducted, and the results show that our model achieves much higher state-of-the-art performance than any single network on Fer2013. In particular, our NA-Resnet achieves 74.59% on Fer2013 and an average accuracy of 96.06% with a standard deviation of 2.9% through 10-fold-cross-validation on Ck+.
引用
收藏
页码:16375 / 16393
页数:19
相关论文
共 46 条
[1]   A novel hidden Markov model-based adaptive dynamic time warping (HMDTW) gait analysis for identifying physically challenged persons [J].
Achanta, Sampath Dakshina Murthy ;
Karthikeyan, T. ;
Vinothkanna, R. .
SOFT COMPUTING, 2019, 23 (18) :8359-8366
[2]  
[Anonymous], 2014, PROC CVPR IEEE, DOI [DOI 10.1109/CVPR.2014.233, 10.1109/CVPR.2014.233]
[3]   Island Loss for Learning Discriminative Features in Facial Expression Recognition [J].
Cai, Jie ;
Meng, Zibo ;
Khan, Ahmed Shehab ;
Li, Zhiyuan ;
O'Reilly, James ;
Tong, Yan .
PROCEEDINGS 2018 13TH IEEE INTERNATIONAL CONFERENCE ON AUTOMATIC FACE & GESTURE RECOGNITION (FG 2018), 2018, :302-309
[4]  
Connie Tee, 2017, Multi-disciplinary Trends in Artificial Intelligence. 11th International Workshop, MIWAI 2017. Proceedings: LNAI 10607, P139, DOI 10.1007/978-3-319-69456-6_12
[5]   Multi-region Ensemble Convolutional Neural Network for Facial Expression Recognition [J].
Fan, Yingruo ;
Lam, Jacqueline C. K. ;
Li, Victor O. K. .
ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2018, PT I, 2018, 11139 :84-94
[6]   Dual Attention Network for Scene Segmentation [J].
Fu, Jun ;
Liu, Jing ;
Tian, Haijie ;
Li, Yong ;
Bao, Yongjun ;
Fang, Zhiwei ;
Lu, Hanqing .
2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, :3141-3149
[7]  
Goodfellow Ian J., 2013, Neural Information Processing. 20th International Conference, ICONIP 2013. Proceedings: LNCS 8228, P117, DOI 10.1007/978-3-642-42051-1_16
[8]   Categorical and dimensional affect analysis in continuous input: Current trends and future directions [J].
Gunes, Hatice ;
Schuller, Bjoern .
IMAGE AND VISION COMPUTING, 2013, 31 (02) :120-136
[9]  
HE KM, 2016, PROC CVPR IEEE, P770, DOI [DOI 10.1109/CVPR.2016.90, 10.1109/CVPR.2016.90]
[10]   Deep Multi-task Learning to Recognise Subtle Facial Expressions of Mental States [J].
Hu, Guosheng ;
Liu, Li ;
Yuan, Yang ;
Yu, Zehao ;
Hua, Yang ;
Zhang, Zhihong ;
Shen, Fumin ;
Shao, Ling ;
Hospedales, Timothy ;
Robertson, Neil ;
Yang, Yongxin .
COMPUTER VISION - ECCV 2018, PT XII, 2018, 11216 :106-123