DISTRIBUTED FOUNDATION MODELS FOR MULTI-MODAL LEARNING IN 6G WIRELESS NETWORKS

被引:42
作者
Du, Jun [1 ]
Lin, Tianyi [2 ]
Jiang, Chunxiao [3 ]
Yang, Qianqian [4 ]
Bader, C. Faouzi [5 ]
Han, Zhu [6 ]
机构
[1] Tsinghua Univ, Commun Engn, Beijing, Peoples R China
[2] Tsinghua Univ, Elect & Commun Engn, Beijing, Peoples R China
[3] Tsinghua Univ, Sch Informat Sci & Technol, Beijing, Peoples R China
[4] Zhejiang Univ, Dept Informat Sci & Elect Engn, Hangzhou, Peoples R China
[5] Technol Innovat Inst, Abu Dhabi, U Arab Emirates
[6] Univ Houston, Elect & Comp Engn Dept, Comp Sci Dept, Houston, TX USA
关键词
COMMUNICATION;
D O I
10.1109/MWC.009.2300501
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Benefiting from the ability to process and integrate data from various modalities, multi-modal foundation models (FMs) facilitate potential applications across a range of fields, including computer vision (CV), natural language processing (NLP), and diverse multi-modal applications such as imagetext retrieval. Currently, FMs are deployed on computing clusters for training and inference to meet their considerable computational demands. In the foreseeable future, the parameter size of FMs is expected to evolve further, posing challenges to both computation resources and energy supply. Fortunately, leveraging the next-generation wireless networks (6G) to aggregate substantial computation resources and multi-modal data from myriad wireless devices holds promise for handling the aforementioned challenges. In this work, we delve into state-of-the-art artificial intelligence (AI) techniques, specifically focusing on pipeline parallelism, data parallelism, and multi-modal learning, with the aim of supporting the sustainable development of distributed multi-modal FMs in the 6G era. In the context of pipeline parallelism, compressing activations and gradients while intelligently allocating communication resources can overcome communication bottlenecks caused by unstable wireless links. For data parallelism, federated learning (FL) with over-the-air computation (AirComp) seamlessly integrates communication and computation, significantly expediting gradient aggregation. Furthermore, by following the recent success of large language models (LLMs) and incorporating multi-modal learning into FMs, we can seamlessly integrate NLP and CV, along with the broader AI community, establishing the cornerstone for the intrinsic AI within 6G wireless networks.
引用
收藏
页码:20 / 30
页数:11
相关论文
共 15 条
[1]   Twelve Scientific Challenges for 6G: Rethinking the Foundations of Communications Theory [J].
Chafii, Marwa ;
Bariah, Lina ;
Muhaidat, Sami ;
Debbah, Merouane .
IEEE COMMUNICATIONS SURVEYS AND TUTORIALS, 2023, 25 (02) :868-904
[2]   Distributed Learning in Wireless Networks: Recent Progress and Future Challenges [J].
Chen, Mingzhe ;
Gunduz, Deniz ;
Huang, Kaibin ;
Saad, Walid ;
Bennis, Mehdi ;
Feljan, Aneta Vulgarakis ;
Poor, H. Vincent .
IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, 2021, 39 (12) :3579-3605
[3]   Gradient and Channel Aware Dynamic Scheduling for Over-the-Air Computation in Federated Edge Learning Systems [J].
Du, Jun ;
Jiang, Bingqing ;
Jiang, Chunxiao ;
Shi, Yuanming ;
Han, Zhu .
IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, 2023, 41 (04) :1035-1050
[4]   Multi-Hop RIS-Empowered Terahertz Communications: A DRL-Based Hybrid Beamforming Design [J].
Huang, Chongwen ;
Yang, Zhaohui ;
Alexandropoulos, George C. ;
Xiong, Kai ;
Wei, Li ;
Yuen, Chau ;
Zhang, Zhaoyang ;
Debbah, Merouane .
IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, 2021, 39 (06) :1663-1677
[5]   Decentralized federated learning through proxy model sharing [J].
Kalra, Shivam ;
Wen, Junfeng ;
Cresswell, Jesse C. ;
Volkovs, Maksims ;
Tizhoosh, H. R. .
NATURE COMMUNICATIONS, 2023, 14 (01)
[6]   Merak: An Efficient Distributed DNN Training Framework With Automated 3D Parallelism for Giant Foundation Models [J].
Lai, Zhiquan ;
Li, Shengwei ;
Tang, Xudong ;
Ge, Keshi ;
Liu, Weijie ;
Duan, Yabo ;
Qiao, Linbo ;
Li, Dongsheng .
IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2023, 34 (05) :1466-1478
[7]   Decentralized Federated Learning: Fundamentals, State of the Art, Frameworks, Trends, and Challenges [J].
Martinez Beltran, Enrique Tomas ;
Perez, Mario Quiles ;
Sanchez, Pedro Miguel Sanchez ;
Bernal, Sergio Lopez ;
Bovet, Gerome ;
Perez, Manuel Gil ;
Perez, Gregorio Martinez ;
Celdran, Alberto Huertas .
IEEE COMMUNICATIONS SURVEYS AND TUTORIALS, 2023, 25 (04) :2983-3013
[8]   Distributed Learning Over Networks With Graph-Attention-Based Personalization [J].
Tian, Zhuojun ;
Zhang, Zhaoyang ;
Yang, Zhaohui ;
Jin, Richeng ;
Dai, Huaiyu .
IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2023, 71 :2071-2086
[9]   Joint Multi-User Communication and Sensing Exploiting Both Signal and Environment Sparsity [J].
Tong, Xin ;
Zhang, Zhaoyang ;
Wang, Jue ;
Huang, Chongwen ;
Debbah, Merouane .
IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2021, 15 (06) :1409-1422
[10]  
Wen D., 2023, IEEE Transactions on Wireless Communications