Dual-ATME: Dual-Branch Attention Network for Micro-Expression Recognition

被引:22
作者
Zhou, Haoliang [1 ,2 ]
Huang, Shucheng [1 ]
Li, Jingting [2 ,3 ]
Wang, Su-Jing [2 ,3 ]
机构
[1] Jiangsu Univ Sci & Technol, Sch Comp, Zhenjiang 212100, Peoples R China
[2] Chinese Acad Sci, Inst Psychol, Key Lab Behav Sci, Beijing 100101, Peoples R China
[3] Univ Chinese Acad Sci, Dept Psychol, Beijing 100049, Peoples R China
基金
中国国家自然科学基金;
关键词
micro-expression recognition; attention mechanism; regions of interest; OPTICAL-FLOW;
D O I
10.3390/e25030460
中图分类号
O4 [物理学];
学科分类号
0702 ;
摘要
Micro-expression recognition (MER) is challenging due to the difficulty of capturing the instantaneous and subtle motion changes of micro-expressions (MEs). Early works based on hand-crafted features extracted from prior knowledge showed some promising results, but have recently been replaced by deep learning methods based on the attention mechanism. However, with limited ME sample sizes, features extracted by these methods lack discriminative ME representations, in yet-to-be improved MER performance. This paper proposes the Dual-branch Attention Network (Dual-ATME) for MER to address the problem of ineffective single-scale features representing MEs. Specifically, Dual-ATME consists of two components: Hand-crafted Attention Region Selection (HARS) and Automated Attention Region Selection (AARS). HARS uses prior knowledge to manually extract features from regions of interest (ROIs). Meanwhile, AARS is based on attention mechanisms and extracts hidden information from data automatically. Finally, through similarity comparison and feature fusion, the dual-scale features could be used to learn ME representations effectively. Experiments on spontaneous ME datasets (including CASME II, SAMM, SMIC) and their composite dataset, MEGC2019-CD, showed that Dual-ATME achieves better, or more competitive, performance than the state-of-the-art MER methods.
引用
收藏
页数:19
相关论文
共 59 条
[1]   Video-Based Facial Micro-Expression Analysis: A Survey of Datasets, Features and Algorithms [J].
Ben, Xianye ;
Ren, Yi ;
Zhang, Junping ;
Wang, Su-Jing ;
Kpalma, Kidiyo ;
Meng, Weixiao ;
Liu, Yong-Jin .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (09) :5826-5846
[2]   Look and Think Twice: Capturing Top-Down Visual Attention with Feedback Convolutional Neural Networks [J].
Cao, Chunshui ;
Liu, Xianming ;
Yang, Yi ;
Yu, Yinan ;
Wang, Jiang ;
Wang, Zilei ;
Huang, Yongzhen ;
Wang, Liang ;
Huang, Chang ;
Xu, Wei ;
Ramanan, Deva ;
Huang, Thomas S. .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, :2956-2964
[3]  
Chaudhry R, 2009, PROC CVPR IEEE, P1932, DOI 10.1109/CVPRW.2009.5206821
[4]   Spatiotemporal Convolutional Neural Network with Convolutional Block Attention Module for Micro-Expression Recognition [J].
Chen, Boyu ;
Zhang, Zhihao ;
Liu, Nian ;
Tan, Yang ;
Liu, Xinyu ;
Chen, Tong .
INFORMATION, 2020, 11 (08)
[5]   Control of goal-directed and stimulus-driven attention in the brain [J].
Corbetta, M ;
Shulman, GL .
NATURE REVIEWS NEUROSCIENCE, 2002, 3 (03) :201-215
[6]   SAMM: A Spontaneous Micro-Facial Movement Dataset [J].
Davison, Adrian K. ;
Lansley, Cliff ;
Costen, Nicholas ;
Tan, Kevin ;
Yap, Moi Hoon .
IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2018, 9 (01) :116-129
[7]  
Dosovitskiy A., 2021, arXiv
[8]  
Ekman P., 2009, Telling lies: Clues to deceit in the marketplace, politics, and marriage, VRevised ed.
[9]   CAS(ME)2: A Database for Spontaneous Macro-Expression and Micro-Expression Spotting and Recognition [J].
Qu, Fangbing ;
Wang, Su-Jing ;
Yan, Wen-Jing ;
Li, He ;
Wu, Shuhang ;
Fu, Xiaolan .
IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2018, 9 (04) :424-436
[10]  
Franke M., 2009, ELECT TECHNOLOGY ISS, P1