Leveraging Contrastive Learning and Self-Training for Multimodal Emotion Recognition with Limited Labeled Samples

被引:1
作者
Fan, Qi [1 ]
Li, Yutong [2 ]
Xin, Yi [3 ]
Cheng, Xinyu [3 ]
Gao, Guanglai [1 ]
Ma, Miao [2 ]
机构
[1] Inner Mongolia Univ, Hohhot, Peoples R China
[2] Shaanxi Normal Univ, Xian, Peoples R China
[3] Nanjing Univ, Nanjing, Peoples R China
来源
PROCEEDINGS OF THE 2ND INTERNATIONAL WORKSHOP ON MULTIMODAL AND RESPONSIBLE AFFECTIVE COMPUTING, MRAC 2024 | 2024年
基金
中国国家自然科学基金;
关键词
Multimodal Emotion Recognition; Semi-Supervised Learning; Contrastive Learning; Multi-Classifier Voting;
D O I
10.1145/3689092.3689412
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The Multimodal Emotion Recognition challenge MER2024 focuses on recognizing emotions using audio, language, and visual signals. In this paper, we present our submission solutions for the Semi-Supervised Learning Sub-Challenge (MER2024-SEMI), which tackles the issue of limited annotated data in emotion recognition. Firstly, to address the class imbalance, we adopt an oversampling strategy. Secondly, we propose a modality representation combinatorial contrastive learning (MR-CCL) framework on the trimodal input data to establish robust initial models. Thirdly, we explore a self-training approach to expand the training set. Finally, we enhance prediction robustness through a multi-classifier weighted soft voting strategy. Our proposed method is validated to be effective on the MER2024-SEMI Challenge, achieving a weighted average F-score of 88.25% and ranking 6th on the leaderboard. Our project is available at https://github.com/WooyoohL/MER2024-SEMI.
引用
收藏
页码:72 / 77
页数:6
相关论文
共 50 条
[21]   A contrastive learning based unsupervised multi-view stereo with multi-stage self-training strategy [J].
Wang, Zihang ;
Luo, Haonan ;
Wang, Xiang ;
Zheng, Jin ;
Ning, Xin ;
Bai, Xiao .
DISPLAYS, 2024, 83
[22]   Self-supervised group meiosis contrastive learning for EEG-based emotion recognition [J].
Haoning Kan ;
Jiale Yu ;
Jiajin Huang ;
Zihe Liu ;
Heqian Wang ;
Haiyan Zhou .
Applied Intelligence, 2023, 53 :27207-27225
[23]   Self-supervised group meiosis contrastive learning for EEG-based emotion recognition [J].
Kan, Haoning ;
Yu, Jiale ;
Huang, Jiajin ;
Liu, Zihe ;
Wang, Heqian ;
Zhou, Haiyan .
APPLIED INTELLIGENCE, 2023, 53 (22) :27207-27225
[24]   Cross-Domain Facial Expression Recognition via Contrastive Warm up and Complexity-Aware Self-Training [J].
Li, Yingjian ;
Huang, Jiaxing ;
Lu, Shijian ;
Zhang, Zheng ;
Lu, Guangming .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2023, 32 :5438-5450
[25]   A contrastive self-supervised learning method for source-free EEG emotion recognition [J].
Wang, Yingdong ;
Ruan, Qunsheng ;
Wu, Qingfeng ;
Wang, Shuocheng .
USER MODELING AND USER-ADAPTED INTERACTION, 2025, 35 (01)
[26]   Comparing Recognition Performance and Robustness of Multimodal Deep Learning Models for Multimodal Emotion Recognition [J].
Liu, Wei ;
Qiu, Jie-Lin ;
Zheng, Wei-Long ;
Lu, Bao-Liang .
IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2022, 14 (02) :715-729
[27]   Cross-modal contrastive learning for multimodal sentiment recognition [J].
Yang, Shanliang ;
Cui, Lichao ;
Wang, Lei ;
Wang, Tao .
APPLIED INTELLIGENCE, 2024, 54 (05) :4260-4276
[28]   Multimodal Named Entity Recognition with Bottleneck Fusion and Contrastive Learning [J].
Wang, Peng ;
Chen, Xiaohang ;
Shang, Ziyu ;
Ke, Wenjun .
IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2023, E106D (04) :545-555
[29]   Modality aware contrastive learning for multimodal human activity recognition [J].
Dixon, Sam ;
Yao, Lina ;
Davidson, Robert .
CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2024, 36 (16)
[30]   Cross-modal contrastive learning for multimodal sentiment recognition [J].
Shanliang Yang ;
Lichao Cui ;
Lei Wang ;
Tao Wang .
Applied Intelligence, 2024, 54 :4260-4276