InMyFace: Inertial and mechanomyography-based sensor fusion for wearable facial activity recognition

被引:14
作者
Bello, Hymalai [1 ]
Marin, Luis Alfredo Sanchez [1 ]
Suh, Sungho [1 ,2 ]
Zhou, Bo [1 ,2 ]
Lukowicz, Paul [1 ,2 ]
机构
[1] German Res Ctr Artificial Intelligence DFKI, D-67663 Kaiserslautern, Germany
[2] RPTU Kaiserslautern Landau, Dept Comp Sci, D-67663 Kaiserslautern, Germany
关键词
Multimodal fusion; Facial expressions; Activity recognition; Mechanomyography; MUSCLE-ACTIVITY; EXPRESSION; PLATFORM;
D O I
10.1016/j.inffus.2023.101886
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recognizing facial activity is a well-understood (but non-trivial) computer vision problem. However, reliable solutions require a camera with a good view of the face, which is often unavailable in wearable settings. Furthermore, in wearable applications, where systems accompany users throughout their daily activities, a permanently running camera can be problematic for privacy (and legal) reasons. This work presents an alternative solution based on the fusion of wearable inertial sensors, planar pressure sensors, and acoustic mechanomyography (muscle sounds). The sensors were placed unobtrusively in a sports cap to monitor facial muscle activities related to facial expressions. We present our integrated wearable sensor system, describe data fusion and analysis methods, and evaluate the system in an experiment with thirteen subjects from different cultural backgrounds (eight countries) and both sexes (six women and seven men). In a one-model-per-user scheme and using a late fusion approach, the system yielded an average F1 score of 85.00% for the case where all sensing modalities are combined. With a cross-user validation and a one-model-for-all-user scheme, an F1 score of 79.00% was obtained for thirteen participants (six females and seven males). Moreover, in a hybrid fusion (cross-user) approach and six classes, an average F1 score of 82.00% was obtained for eight users. The results are competitive with state-of-the-art non-camera-based solutions for a cross-user study. In addition, our unique set of participants and minimally biased experimental design demonstrate the inclusiveness of the approach, which is beneficial for further generalizability.
引用
收藏
页数:14
相关论文
共 87 条
[21]   Hierarchical Classification and Transfer Learning to Recognize Head Gestures and Facial Expressions Using Earbuds [J].
Gashi, Shkurta ;
Saeed, Aaqib ;
Vicini, Alessandra ;
Di Lascio, Elena ;
Santini, Silvia .
PROCEEDINGS OF THE 2021 INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION, ICMI 2021, 2021, :168-176
[22]   Wearable Body Sensor Networks: State-of-the-Art and Research Directions [J].
Gravina, Raffaele ;
Fortino, Giancarlo .
IEEE SENSORS JOURNAL, 2021, 21 (11) :12511-12522
[23]   Mechanomyography Assisted Myoeletric Sensing for Upper-Extremity Prostheses: A Hybrid Approach [J].
Guo, Weichao ;
Sheng, Xinjun ;
Liu, Honghai ;
Zhu, Xiangyang .
IEEE SENSORS JOURNAL, 2017, 17 (10) :3100-3108
[24]  
Guo X., 2023, IEEE INTERNET THINGS
[25]  
Harper R.G., 1978, NONVERBAL COMMUNICAT
[26]  
Hasan M.R., 2004, VARIATIONS, V1, P4
[27]   Eyemotion: Classifying facial expressions in VR using eye-tracking cameras [J].
Hickson, Steven ;
Dufour, Nick ;
Sud, Avneesh ;
Kwatra, Vivek ;
Essa, Irfan .
2019 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2019, :1626-1635
[28]  
Hossain M.R., 2010, 2010 4 INT C SIGNAL, P1, DOI [DOI 10.1109/ICSPCS.2010.5709752, 10.1109/ICSPCS.2010.5709752]
[29]   Face2Multi-modal: In-vehicle Multi-modal Predictors via Facial Expressions [J].
Huang, Zhentao ;
Li, Rongze ;
Jin, Wangkai ;
Song, Zilin ;
Zhang, Yu ;
Peng, Xiangjun ;
Sun, Xu .
12TH INTERNATIONAL ACM CONFERENCE ON AUTOMOTIVE USER INTERFACES AND INTERACTIVE VEHICULAR APPLICATIONS, AUTOMOTIVEUI 2020, 2020, :30-33
[30]   Affective Information in Context and Judgment of Facial Expression: Cultural Similarities and Variations in Context Effects Between North Americans and East Asians [J].
Ito, Kenichi ;
Masuda, Takahiko ;
Hioki, Koichi .
JOURNAL OF CROSS-CULTURAL PSYCHOLOGY, 2012, 43 (03) :429-445