EmoComicNet: A multi-task model for comic emotion recognition

被引:4
|
作者
Dutta, Arpita [1 ,2 ]
Biswas, Samit [1 ]
Das, Amit Kumar [1 ]
机构
[1] Indian Inst Engn Science&Technol, Dept Comp Science&Technol, Howrah 711103, West Bengal, India
[2] Techno Main, Artificial Intelligence & Machine Learning, Dept Comp Sci & Engn, Kolkata 700091, West Bengal, India
关键词
Comic analysis; Multi-modal emotion recognition; Document image processing; Deep learning; Multi-task learning;
D O I
10.1016/j.patcog.2024.110261
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The emotion and sentiment associated with comic scenes can provide potential information for inferring the context of comic stories, which is an essential pre -requisite for developing comics' automatic content understanding tools. Here, we address this open area of comic research by exploiting the multi -modal nature of comics. The general assumptions for multi -modal sentiment analysis methods are that both image and text modalities are always present at the test phase. However, this assumption is not always satisfied for comics since comic characters' facial expressions, gestures, etc., are not always clearly visible. Also, the dialogues between comic characters are often challenging to comprehend the underlying context. To deal with these constraints of comic emotion analysis, we propose a multi -task -based framework, namely EmoComicNet, to fuse multi -modal information (i.e., both image and text) if it is available. However, the proposed EmoComicNet is designed to perform even when any modality is weak or completely missing. The proposed method potentially improves the overall performance. Besides, EmoComicNet can also deal with the problem of weak or absent modality during the training phase.
引用
收藏
页数:11
相关论文
共 50 条
  • [21] Speech Emotion Recognition using Decomposed Speech via Multi-task Learning
    Hsu, Jia-Hao
    Wu, Chung-Hsien
    Wei, Yu-Hung
    INTERSPEECH 2023, 2023, : 4553 - 4557
  • [22] Coarse-to-Fine Speech Emotion Recognition Based on Multi-Task Learning
    Zhao, Huijuan
    Ye, Ning
    Wang, Ruchuan
    JOURNAL OF SIGNAL PROCESSING SYSTEMS FOR SIGNAL IMAGE AND VIDEO TECHNOLOGY, 2021, 93 (2-3): : 299 - 308
  • [23] Multi-Task Semi-Supervised Adversarial Autoencoding for Speech Emotion Recognition
    Latif, Siddique
    Rana, Rajib
    Khalifa, Sara
    Jurdak, Raja
    Epps, Julien
    Schuller, Bjoern W.
    IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2022, 13 (02) : 992 - 1004
  • [24] Multi-modal Sentiment and Emotion Joint Analysis with a Deep Attentive Multi-task Learning Model
    Zhang, Yazhou
    Rong, Lu
    Li, Xiang
    Chen, Rui
    ADVANCES IN INFORMATION RETRIEVAL, PT I, 2022, 13185 : 518 - 532
  • [25] Multi-label, multi-task CNN approach for context-based emotion recognition
    Bendjoudi, Ilyes
    Vanderhaegen, Frederic
    Hamad, Denis
    Dornaika, Fadi
    INFORMATION FUSION, 2021, 76 : 422 - 428
  • [26] FasterPest: A Multi-Task Classification Model for Rice Pest Recognition
    Zhan, Xiaoyun
    Zhang, Cong
    Wang, Zheng
    Han, Yuantao
    Xiong, Peng
    He, Linfeng
    IEEE ACCESS, 2024, 12 : 167845 - 167855
  • [27] FasterPest: A Multi-Task Classification Model for Rice Pest Recognition
    Zhan, Xiaoyun
    Zhang, Cong
    Wang, Zheng
    Han, Yuantao
    Xiong, Peng
    He, Linfeng
    IEEE ACCESS, 2024, 12 : 167845 - 167855
  • [28] IMPROVING SAR TARGET RECOGNITION WITH MULTI-TASK LEARNING
    Du, Wenrui
    Zhang, Fan
    Ma, Fei
    Yin, Qiang
    Zhou, Yongsheng
    IGARSS 2020 - 2020 IEEE INTERNATIONAL GEOSCIENCE AND REMOTE SENSING SYMPOSIUM, 2020, : 284 - 287
  • [29] A Co-regularization Facial Emotion Recognition Based on Multi-Task Facial Action Unit Recognition
    Udeh, Chinonso Paschal
    Chen, Luefeng
    Du, Sheng
    Li, Min
    Wu, Min
    2022 41ST CHINESE CONTROL CONFERENCE (CCC), 2022, : 6806 - 6810
  • [30] A Multi-Task Framework for Weather Recognition
    Li, Xuelong
    Wang, Zhigang
    Lu, Xiaoqiang
    PROCEEDINGS OF THE 2017 ACM MULTIMEDIA CONFERENCE (MM'17), 2017, : 1318 - 1326