Triple-ATFME: Triple-Branch Attention Fusion Network for Micro-Expression Recognition

被引:4
作者
Li, Fei [1 ]
Nie, Ping [1 ]
You, Meiming [1 ]
Chen, Zhichao [1 ]
Wang, Guoqiang [1 ]
机构
[1] Heilongjiang Univ, Coll Elect Engn, Harbin, Peoples R China
关键词
Micro-expression recognition; Optical flow; Composite database; Attention mechanism; DECEPTION;
D O I
10.1007/s13369-024-08973-z
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Micro-expressions (MEs) are momentary and subtle facial expression changes that typically have a short duration to reveal people's genuine emotions and thought processes. With the continuous advancement of deep learning techniques, more opportunities and challenges have been presented for research in the field of micro-expression recognition (MER). Due to the limitations and low intensity of ME movements, as well as the presence of subtle motions and limited sample data, a single-stream model structure that extracts features from a single view is unable to capture sufficient information. Therefore, this paper proposes the Triple-branch attention fusion network (Triple-ATFME) for MER. The proposed Triple-ATFME mainly consists of a preprocessing stage, a Triple-branch ShuffleNet module, and an adaptive channel attention module allowing the model to extract multi-view features through a multi-path network. Firstly, the MER framework extracts optical flow features from the cropped initial frame and peak frame of the facial region using the optical flow method. Multiple types of optical flow features are obtained. Secondly, the extracted features are put into the Triple-ATFME network to extract deep-level hidden features. The three sub-modules extract multi-view features. In order to alleviate the issue of the tendency that the model excessively focuses on local information during the feature fusion process, a simple effective adaptive channel fusion attention module (CFAM) is employed to calibrate the channel features and to facilitate the fusion of multi-view features. Finally, ablation experiments are conducted to demonstrate that the multi-view features and the Triple-ATFME network are more effective in learning MEs. Experimental results show that the proposed Triple-ATFME achieves UF1 and UAR of 0.7609 and 0.7565 on the combined dataset MEGC2019, demonstrating superior performance compared to state-of-the-art MER methods as evaluated in leave-one-out validation.
引用
收藏
页码:807 / 823
页数:17
相关论文
共 48 条
[1]   Block Division Convolutional Network With Implicit Deep Features Augmentation for Micro-Expression Recognition [J].
Chen, Bin ;
Liu, Kun-Hong ;
Xu, Yong ;
Wu, Qing-Qiang ;
Yao, Jun-Feng .
IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 :1345-1358
[2]   SAMM: A Spontaneous Micro-Facial Movement Dataset [J].
Davison, Adrian K. ;
Lansley, Cliff ;
Costen, Nicholas ;
Tan, Kevin ;
Yap, Moi Hoon .
IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2018, 9 (01) :116-129
[3]  
Davison S.T., 2019, ARXIV
[4]   Darwin, deception, and facial expression [J].
Ekman, P .
EMOTIONS INSIDE OUT: 130 YEARS AFTER DARWIN'S THE EXPRESSION OF THE EMOTIONS IN MAN AND ANIMALS, 2003, 1000 :205-221
[5]  
Ekman P., 2009, Telling Lies: Clues to Deceit in the Marketplace, Politics, and Marriage, DOI DOI 10.1080/00029157.2011.10404358
[6]   OFF-ApexNet on micro-expression recognition system [J].
Gan, Y. S. ;
Liong, Sze-Teng ;
Yau, Wei-Chuen ;
Huang, Yen-Chang ;
Tan, Lit-Ken .
SIGNAL PROCESSING-IMAGE COMMUNICATION, 2019, 74 :129-139
[7]   Fuzzy Histogram of Optical Flow Orientations for Micro-Expression Recognition [J].
Happy, S. L. ;
Routray, Aurobinda .
IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2019, 10 (03) :394-406
[8]   DETERMINING OPTICAL-FLOW [J].
HORN, BKP ;
SCHUNCK, BG .
ARTIFICIAL INTELLIGENCE, 1981, 17 (1-3) :185-203
[9]  
Hu J, 2018, PROC CVPR IEEE, P7132, DOI [10.1109/CVPR.2018.00745, 10.1109/TPAMI.2019.2913372]
[10]   Spontaneous facial micro-expression analysis using Spatiotemporal Completed Local Quantized Patterns [J].
Huang, Xiaohua ;
Zhao, Guoying ;
Hong, Xiaopeng ;
Zheng, Wenming ;
Pietikainen, Matti .
NEUROCOMPUTING, 2016, 175 :564-578