Sparse landmarks for facial action unit detection using vision transformer and perceiver

被引:0
作者
Cakir, Duygu [1 ]
Yilmaz, Gorkem [2 ]
Arica, Nafiz [3 ]
机构
[1] Bahcesehir Univ, Fac Engn & Nat Sci, Dept Software Engn, Istanbul, Turkiye
[2] Bahcesehir Univ, Fac Engn & Nat Sci, Dept Comp Engn, Istanbul, Turkiye
[3] Piri Reis Univ, Fac Engn, Dept Informat Syst Engn, Istanbul, Turkiye
关键词
action unit detection; sparse learning; vision transformer; perceiver; RECOGNITION; PATCHES;
D O I
10.1504/IJCSE.2023.10060451
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
The ability to accurately detect facial expressions, represented by facial action units (AUs), holds significant implications across diverse fields such as mental health diagnosis, security, and human-computer interaction. Although earlier approaches have made progress, the burgeoning complexity of facial actions demands more nuanced, computationally efficient techniques. This study pioneers the integration of sparse learning with vision transformer (ViT) and perceiver networks, focusing on the most active and descriptive landmarks for AU detection across both controlled (DISFA, BP4D) and in-the-wild (EmotioNet) datasets. Our novel approach, employing active landmark patches instead of the whole face, not only attains state-of-the-art performance but also uncovers insights into the differing attention mechanisms of ViT and perceiver. This fusion of techniques marks a significant advancement in facial analysis, potentially reshaping strategies in noise reduction and patch optimisation, setting a robust foundation for future research in the domain.
引用
收藏
页码:607 / 620
页数:15
相关论文
共 50 条
  • [31] COVID-Transformer: Interpretable COVID-19 Detection Using Vision Transformer for Healthcare
    Shome, Debaditya
    Kar, T.
    Mohanty, Sachi Nandan
    Tiwari, Prayag
    Muhammad, Khan
    AlTameem, Abdullah
    Zhang, Yazhou
    Saudagar, Abdul Khader Jilani
    INTERNATIONAL JOURNAL OF ENVIRONMENTAL RESEARCH AND PUBLIC HEALTH, 2021, 18 (21)
  • [32] Comparison of Eye-gaze Detection using CNN and Vision Transformer
    Niikura D.
    Abe K.
    IEEJ Transactions on Electronics, Information and Systems, 2024, 144 (07) : 683 - 684
  • [33] CrimeNet: Neural Structured Learning using Vision Transformer for violence detection
    Rendon-Segador, Fernando J.
    Alvarez-Garcia, Juan A.
    Salazar-Gonzalez, Jose L.
    Tommasi, Tatiana
    NEURAL NETWORKS, 2023, 161 : 318 - 329
  • [34] Anomaly detection in cropland monitoring using multiple view vision transformer
    Xuesong Liu
    Yansong Liu
    He Sui
    Chuan Qin
    Yuanxi Che
    Zhaobo Guo
    Scientific Reports, 15 (1)
  • [35] Comparative Analysis of Vision Transformer Models for Facial Emotion Recognition Using Augmented Balanced Datasets
    Bobojanov, Sukhrob
    Kim, Byeong Man
    Arabboev, Mukhriddin
    Begmatov, Shohruh
    APPLIED SCIENCES-BASEL, 2023, 13 (22):
  • [36] LOCAL NORMAL BINARY PATTERNS FOR 3D FACIAL ACTION UNIT DETECTION
    Sandbach, Georgia
    Zafeiriou, Stefanos
    Pantic, Maja
    2012 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP 2012), 2012, : 1813 - 1816
  • [37] Electromyographic Validation of Spontaneous Facial Mimicry Detection Using Automated Facial Action Coding
    Hsu, Chun-Ting
    Sato, Wataru
    SENSORS, 2023, 23 (22)
  • [38] Multi-conditional Latent Variable Model for Joint Facial Action Unit Detection
    Eleftheriadis, Stefanos
    Rudovic, Ognjen
    Pantic, Maja
    2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, : 3792 - 3800
  • [39] Facial Action Unit Detection and Intensity Estimation From Self-Supervised Representation
    Ma, Bowen
    An, Rudong
    Zhang, Wei
    Ding, Yu
    Zhao, Zeng
    Zhang, Rongsheng
    Lv, Tangjie
    Fan, Changjie
    Hu, Zhipeng
    IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2024, 15 (03) : 1669 - 1683
  • [40] ViTSigat: Early Black Sigatoka Detection in Banana Plants Using Vision Transformer
    Charco, Jorge L.
    Yanza-Montalvan, Angela
    Zumba-Gamboa, Johanna
    Alonso-Anguizaca, Jose
    Basurto-Cruz, Edgar
    INFORMATION AND COMMUNICATION TECHNOLOGIES, TICEC 2024, 2025, 2273 : 117 - 130