Robust multimodal federated learning for incomplete modalities

被引:3
|
作者
Yu, Songcan [1 ,2 ]
Wang, Junbo [1 ]
Hussein, Walid [3 ]
Hung, Patrick C. K. [4 ]
机构
[1] Sun Yat Sen Univ, Sch Intelligent Syst Engn, Shenzhen 518107, Peoples R China
[2] Guangdong Prov Key Lab Fire Sci & Intelligent Emer, Guangzhou 510006, Peoples R China
[3] British Univ Egypt, Fac Informat & Comp Sci, Cairo, Egypt
[4] Ontario Tech Univ, Fac Business & Informat Technol, Oshawa, ON, Canada
关键词
Multimodal fusion; Federated learning; Data incompleteness; Missing modalities;
D O I
10.1016/j.comcom.2023.12.003
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Consumer electronics are continuously collecting multimodal data, such as audio, video, and so on. A multimodal learning mechanism can be adopted to deal with these data. Due to the consideration of privacy protection, some successful attempts at multimodal federated learning (MMFed) have been conducted. However, real-world multimodal data is usually missing modalities, which can significantly affect the accuracy of the global model in MMFed. Effectively fusing and analyzing multimodal data with incompleteness remains a challenging problem. To tackle this problem, we propose a robust Multimodal Federated Learning Framework for Incomplete Modalities (FedInMM). More specifically, we design a Long Short-Term Memory (LSTM)-based module to extract the information in the temporal sequence. We dynamically learn a weight map to rescale the feature in each modality and formulate the different contributions of features. And then the content of each modality is further fused to form a uniform representation of all modalities of data. By considering the temporal dependency and intra-relation of multi-modalities automatically through the learning stage, this MMFed framework can efficiently mitigate the effects of missing modalities. By using two multimodal datasets, DEAP and AReM, we have conducted comprehensive experiments by simulating different levels of incompleteness. Experimental results demonstrate that FedInMM outperforms other approaches and can train highly accurate models on datasets comprising different incompleteness patterns, which is more appropriate for integration into a practical multimodal application.
引用
收藏
页码:234 / 243
页数:10
相关论文
共 50 条
  • [1] A Multimodal Federated Learning Framework for Modality Incomplete Scenarios in Healthcare
    An, Ying
    Bai, Yaqi
    Liu, Yuan
    Guo, Lin
    Chen, Xianlai
    BIOINFORMATICS RESEARCH AND APPLICATIONS, PT II, ISBRA 2024, 2024, 14955 : 245 - 256
  • [2] Chameleon: A Multimodal Learning Framework Robust to Missing Modalities
    Muhammad Irzam Liaqat
    Shah Nawaz
    Muhammad Zaigham Zaheer
    Muhammad Saad Saeed
    Hassan Sajjad
    Tom De Schepper
    Karthik Nandakumar
    Muhammad Haris Khan
    Ignazio Gallo
    Markus Schedl
    International Journal of Multimedia Information Retrieval, 2025, 14 (2)
  • [3] Robust Privacy-Preserving Recommendation Systems Driven by Multimodal Federated Learning
    Feng, Chenyuan
    Feng, Daquan
    Huang, Guanxin
    Liu, Zuozhu
    Wang, Zhenzhong
    Xia, Xiang-Gen
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, : 1 - 15
  • [4] Byzantine-Robust Multimodal Federated Learning Framework for Intelligent Connected Vehicle
    Wu, Ning
    Lin, Xiaoming
    Lu, Jianbin
    Zhang, Fan
    Chen, Weidong
    Tang, Jianlin
    Xiao, Jing
    ELECTRONICS, 2024, 13 (18)
  • [5] AutoFed: Heterogeneity-Aware Federated Multimodal Learning for Robust Autonomous Driving
    Zheng, Tianyue
    Li, Ang
    Chen, Zhe
    Wang, Hongbo
    Luo, Jun
    PROCEEDINGS OF THE 29TH ANNUAL INTERNATIONAL CONFERENCE ON MOBILE COMPUTING AND NETWORKING, MOBICOM 2023, 2023, : 209 - 223
  • [6] Multimodal Federated Learning: A Survey
    Che, Liwei
    Wang, Jiaqi
    Zhou, Yao
    Ma, Fenglong
    SENSORS, 2023, 23 (15)
  • [7] FedMultimodal: A Benchmark For Multimodal Federated Learning
    Feng, Tiantian
    Bose, Digbalay
    Zhang, Tuo
    Hebbar, Rajat
    Ramakrishna, Anil
    Gupta, Rahul
    Zhang, Mi
    Avestimehr, Salman
    Narayanan, Shrikanth
    PROCEEDINGS OF THE 29TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2023, 2023, : 4035 - 4045
  • [8] Robust Aggregation for Federated Learning
    Pillutla, Krishna
    Kakade, Sham M.
    Harchaoui, Zaid
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2022, 70 : 1142 - 1154
  • [9] Multimodal federated learning: Concept, methods, applications and future directions
    Huang, Wei
    Wang, Dexian
    Ouyang, Xiaocao
    Wan, Jihong
    Liu, Jia
    Li, Tianrui
    INFORMATION FUSION, 2024, 112
  • [10] Robust Multimodal Representation under Uncertain Missing Modalities
    Lan, Guilin
    Du, Yeqian
    Yang, Zhouwang
    ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2025, 21 (01)