Mixture of Experts for Intelligent Networks: A Large Language Model-enabled Approach

被引:0
|
作者
Du, Hongyang [1 ]
Liu, Guangyuan [1 ]
Lin, Yijing [2 ]
Niyato, Dusit [1 ]
Kang, Jiawen [3 ,4 ,5 ]
Xiong, Zehui [6 ]
Kim, Dong In [7 ]
机构
[1] Nanyang Technol Univ, Sch Comp Sci & Engn, Singapore 639798, Singapore
[2] Beijing Univ Posts & Telecommun, State Key Lab Networking & Switching Technol, Beijing 100876, Peoples R China
[3] Guangdong Univ Technol, Sch Automat, Guangzhou 510006, Peoples R China
[4] Minist Educ, Key Lab Intelligent Informat Proc & Syst Integrat, Guangzhou 510006, Peoples R China
[5] Guangdong HongKong Macao Joint Lab Smart Discrete, Guangzhou 510006, Peoples R China
[6] Singapore Univ Technol & Design, Pillar Informat Syst Technol & Design, Singapore 487372, Singapore
[7] Sungkyunkwan Univ, Dept Elect & Comp Engn, Suwon 16419, South Korea
来源
20TH INTERNATIONAL WIRELESS COMMUNICATIONS & MOBILE COMPUTING CONFERENCE, IWCMC 2024 | 2024年
基金
中国国家自然科学基金; 新加坡国家研究基金会;
关键词
Generative AI (GAI); large language model; mixture of experts; network optimization;
D O I
10.1109/IWCMC61514.2024.10592370
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Optimizing various wireless user tasks poses a significant challenge for networking systems because of the expanding range of user requirements. Despite advancements in Deep Reinforcement Learning (DRL), the need for customized optimization tasks for individual users complicates developing and applying numerous DRL models, leading to substantial computation resource and energy consumption and can lead to inconsistent outcomes. To address this issue, we propose a novel approach utilizing a Mixture of Experts (MoE) framework, augmented with Large Language Models (LLMs), to analyze user objectives and constraints effectively, select specialized DRL experts, and weigh each decision from the participating experts. Specifically, we develop a gate network to oversee the expert models, allowing a collective of experts to tackle a wide array of new tasks. Furthermore, we innovatively substitute the traditional gate network with an LLM, leveraging its advanced reasoning capabilities to manage expert model selection for joint decisions. Our proposed method reduces the need to train new DRL models for each unique optimization problem, decreasing energy consumption and AI model implementation costs. The LLM-enabled MoE approach is validated through a general maze navigation task and a specific network service provider utility maximization task, demonstrating its effectiveness and practical applicability in optimizing complex networking systems.
引用
收藏
页码:531 / 536
页数:6
相关论文
共 49 条
  • [31] WorkloadGPT: A Large Language Model Approach to Real-Time Detection of Pilot Workload
    Gao, Yijing
    Yue, Lishengsa
    Sun, Jiahang
    Shan, Xiaonian
    Liu, Yihan
    Wu, Xuerui
    APPLIED SCIENCES-BASEL, 2024, 14 (18):
  • [32] An improved transformer-based model for detecting phishing, spam and ham emails: A large language model approach
    Jamal, Suhaima
    Wimmer, Hayden
    Sarker, Iqbal H.
    SECURITY AND PRIVACY, 2024, 7 (05)
  • [33] More intelligent knowledge graph: A large language model-driven method for knowledge representation in geospatial digital twins
    Zhang, Jinbin
    Zhu, Jun
    Guo, Zhihao
    Wu, Jianlin
    Guo, Yukun
    Lai, Jianbo
    Li, Weilian
    INTERNATIONAL JOURNAL OF APPLIED EARTH OBSERVATION AND GEOINFORMATION, 2025, 139
  • [34] Nadine: A large language model-driven intelligent social robot with affective capabilities and human-like memory
    Kang, Hangyeol
    Ben Moussa, Maher
    Thalmann, Nadia Magnenat
    COMPUTER ANIMATION AND VIRTUAL WORLDS, 2024, 35 (04)
  • [35] Evaluating the quality of student-generated content in learnersourcing: A large language model based approach
    Li, Kangkang
    Qian, Chengyang
    Yang, Xianmin
    EDUCATION AND INFORMATION TECHNOLOGIES, 2025, 30 (02) : 2331 - 2360
  • [36] Retrieval-Augmented Generation Approach: Document Question Answering using Large Language Model
    Muludi, Kurnia
    Fitria, Kaira Milani
    Triloka, Joko
    Sutedi
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2024, 15 (03) : 776 - 785
  • [37] Identifying Rare Circumstances Preceding Female Firearm Suicides: Validating A Large Language Model Approach
    Zhou, Weipeng
    Prater, Laura C.
    Goldstein, Evan, V
    Mooney, Stephen J.
    JMIR MENTAL HEALTH, 2023, 10
  • [38] Large language model-based approach for human-mobile inspection robot interactive navigation
    Wang T.
    Fan J.
    Zheng P.
    Jisuanji Jicheng Zhizao Xitong/Computer Integrated Manufacturing Systems, CIMS, 2024, 30 (05): : 1587 - 1594
  • [39] Thinking Like an Author: A Zero-Shot Learning Approach to Keyphrase Generation with Large Language Model
    Wang, Siyu
    Dai, Shengran
    Jiang, Jianhui
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES: RESEARCH TRACK, PT III, ECML PKDD 2024, 2024, 14943 : 335 - 350
  • [40] Engaging Preference Optimization Alignment in Large Language Model for Continual Radiology Report Generation: A Hybrid Approach
    Izhar, Amaan
    Idris, Norisma
    Japar, Nurul
    COGNITIVE COMPUTATION, 2025, 17 (01)