Global-Local combined features to detect pain intensity from facial expression images with attention mechanism

被引:1
作者
Wu, Jiang [1 ,2 ]
Shi, Yi [2 ]
Yan, Shun [3 ]
Yan, Hong-mei [1 ,2 ]
机构
[1] Yangtze Delta Region Institute (Huzhou), University of Electronic Science and Technology of China, Huzhou
[2] MOE Key Laboratory for Neuroinformation, School of Life Science and Technology, University of Electronic Science and Technology of China, Chengdu
[3] College of Engineering, University of California at Santa Barbara, Santa Barbara
基金
中国国家自然科学基金;
关键词
Attention; Convolutional neural network; Facial expression; Pain intensity;
D O I
10.1016/j.jnlest.2024.100260
中图分类号
学科分类号
摘要
The estimation of pain intensity is critical for medical diagnosis and treatment of patients. With the development of image monitoring technology and artificial intelligence, automatic pain assessment based on facial expression and behavioral analysis shows a potential value in clinical applications. This paper reports a framework of convolutional neural network with global and local attention mechanism (GLA-CNN) for the effective detection of pain intensity at four-level thresholds using facial expression images. GLA-CNN includes two modules, namely global attention network (GANet) and local attention network (LANet). LANet is responsible for extracting representative local patch features of faces, while GANet extracts whole facial features to compensate for the ignored correlative features between patches. In the end, the global correlational and local subtle features are fused for the final estimation of pain intensity. Experiments under the UNBC-McMaster Shoulder Pain database demonstrate that GLA-CNN outperforms other state-of-the-art methods. Additionally, a visualization analysis is conducted to present the feature map of GLA-CNN, intuitively showing that it can extract not only local pain features but also global correlative facial ones. Our study demonstrates that pain assessment based on facial expression is a non-invasive and feasible method, and can be employed as an auxiliary pain assessment tool in clinical practice. © 2024 The Authors
引用
收藏
相关论文
共 49 条
[31]  
Liu H., Zheng C., Li D.-T.-C., Et al., Multi-perspective social recommendation method with graph representation learning, Neurocomputing, 468, pp. 469-481, (2022)
[32]  
Liu H., Zhang C., Deng Y.-J., Xie B.-C., Liu T.-T., Li Y.-F., TransIFC: invariant cues-aware feature concentration learning for efficient fine-grained bird image classification, IEEE T. Multimedia, (2023)
[33]  
Liu H., Fang S., Zhang Z.-L., Li D.-T.-C.H., Lin K., Wang J.-Z.H., MFDNet: Collaborative poses perception and matrixfisher distribution for head pose estimation, IEEE T. on Multimedia, 24, pp. 2449-2460, (2021)
[34]  
Li Y., Zeng J.-B., Shan S.-G., Chen X.-L., Occlusion aware facial expression recognition using CNN with attentionmechanism, IEEE T. Image Process, 28, 5, pp. 2439-2450, (2019)
[35]  
Zhang H., Su W., Wang Z., Weakly supervised local-global attention network for facial expression recognition, 8, pp. 37976-37987, (2020)
[36]  
Liu C., Hirota K., Ma J., Et al., Facial expression recognition using hybrid features of pixel and geometry, IEEE Access, 9, pp. 18876-18889, (2021)
[37]  
Liu X., Li J., Duan T.-T., Li J.-T., Wang Y., DSMA: reference-based image super-resolution method based on dual-view supervised learning and multi-attention mechanism, IEEE Access, 10, pp. 54649-54659, (2022)
[38]  
Jaderberg M., Simonyan K., Zisserman A., Kavukcuoglu K., Spatial transformer networks, Proc. of the 28th Intl. Conf. on Neural Information Processing Systems, pp. 2017-2025, (2015)
[39]  
Zhang K.-P., Zhang Z.-P., Li Z.-F., Qian Y., Joint face detection and alignment using multitask cascaded convolutional networks, IEEE Signal Proc. Let., 23, 10, pp. 1499-1503, (2016)
[40]  
Benarroch E.E., Pain-autonomic interactions: a selective review, Clin. Auton. Res., 11, 6, pp. 343-349, (2001)