Enhancing Feature Correlation for Bi-Modal Group Emotion Recognition

被引:6
|
作者
Liu, Ningjie [1 ]
Fang, Yuchun [1 ]
Guo, Yike [1 ,2 ]
机构
[1] Shanghai Univ, Sch Comp Engn & Sci, Shanghai, Peoples R China
[2] Imperial Coll London, Dept Comp, London, England
基金
中国国家自然科学基金;
关键词
Group emotion recognition; B-CNN; Non-local block; AROUSAL;
D O I
10.1007/978-3-030-00767-6_3
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Group emotion recognition in the wild has received much attention in computer vision community. It is a very challenge issue, due to interactions taking place between various numbers of people, different occlusions. According to human cognitive and behavioral researches, background and facial expression play a dominating role in the perception of group's mood. Hence, in this paper, we propose a novel approach that combined these two features for image-based group emotion recognition with feature correlation enhancement. The feature enhancement is mainly reflected in two parts. For facial expression feature extraction, we plug non-local blocks into Xception network to enhance the feature correlation of different positions in low-level, which can avoid the fast loss of position information of the traditional CNNs and effectively enhance the network's feature representation capability. For global scene information, we build a bilinear convolutional neural network (B-CNN) consisting of VGG16 networks to model local pairwise feature interactions in a translationally invariant manner. The experimental results show that the fused feature could effectively improve the performance.
引用
收藏
页码:24 / 34
页数:11
相关论文
共 50 条
  • [21] Bi-modal Handwritten Text Recognition (BiHTR) ICPR 2010 Contest Report
    Pastor, Moises
    Paredes, Roberto
    RECOGNIZING PATTERNS IN SIGNALS, SPEECH, IMAGES, AND VIDEOS, 2010, 6388 : 1 - 13
  • [22] A bi-modal face recognition framework integrating facial expression with facial appearance
    Tsai, Pohsiang
    Cao, Longbing
    Hintz, Tom
    Jan, Tony
    PATTERN RECOGNITION LETTERS, 2009, 30 (12) : 1096 - 1109
  • [23] Bi-Modal Person Recognition on a Mobile Phone: using mobile phone data
    McCool, Chris
    Marcel, Sebastien
    Hadid, Abdenour
    Pietikainen, Matti
    Matejka, Pavel
    Cernocky, Jan
    Poh, Norman
    Kittler, Josef
    Larcher, Anthony
    Levy, Christophe
    Matrouf, Driss
    Bonastre, Jean-Francois
    Tresadern, Phil
    Cootes, Timothy
    2012 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO WORKSHOPS (ICMEW), 2012, : 635 - 640
  • [24] Bi-modal strategy of gastrulation in reptiles
    Stower, Matthew J.
    Diaz, Raul E.
    Fernandez, Lucia Carrera
    Crother, Mary White
    Crother, Brian
    Marco, Adolfo
    Trainor, Paul A.
    Srinivas, Shankar
    Bertocchini, Federica
    DEVELOPMENTAL DYNAMICS, 2015, 244 (09) : 1144 - 1157
  • [25] BI-MODAL NAIVE SET THEORY
    Wigglesworth, John
    AUSTRALASIAN JOURNAL OF LOGIC, 2018, 15 (02) : 139 - 150
  • [26] Strength and Ductility of Bi-Modal Cu
    Zhao, Yonghao
    Topping, Troy
    Li, Ying
    Lavernia, Enrique J.
    ADVANCED ENGINEERING MATERIALS, 2011, 13 (09) : 865 - 871
  • [27] A Survey of Face Recognition Techniques and Comparative Study of Various Bi-Modal and Multi-Modal Techniques
    Handa, Anand
    Agarwal, Rashi
    Kohli, Narendra
    2016 11TH INTERNATIONAL CONFERENCE ON INDUSTRIAL AND INFORMATION SYSTEMS (ICIIS), 2016, : 274 - 279
  • [28] A sketch recognition method based on bi-modal model using cooperative learning paradigm
    Zhang S.
    Wang L.
    Cui Z.
    Wang S.
    Neural Computing and Applications, 2024, 36 (23) : 14275 - 14290
  • [29] ON UNI-MODAL AND BI-MODAL ANNUAL RHYTHMS IN MAN
    KLINKER, L
    JOURNAL OF INTERDISCIPLINARY CYCLE RESEARCH, 1985, 16 (04): : 276 - 276
  • [30] State synchronous modeling of audio-visual information for bi-modal speech recognition
    Nakamura, S
    Kumatani, K
    Tamura, S
    ASRU 2001: IEEE WORKSHOP ON AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING, CONFERENCE PROCEEDINGS, 2001, : 409 - 412