Mixture of Experts for Intelligent Networks: A Large Language Model-enabled Approach

被引:0
|
作者
Du, Hongyang [1 ]
Liu, Guangyuan [1 ]
Lin, Yijing [2 ]
Niyato, Dusit [1 ]
Kang, Jiawen [3 ,4 ,5 ]
Xiong, Zehui [6 ]
Kim, Dong In [7 ]
机构
[1] Nanyang Technol Univ, Sch Comp Sci & Engn, Singapore 639798, Singapore
[2] Beijing Univ Posts & Telecommun, State Key Lab Networking & Switching Technol, Beijing 100876, Peoples R China
[3] Guangdong Univ Technol, Sch Automat, Guangzhou 510006, Peoples R China
[4] Minist Educ, Key Lab Intelligent Informat Proc & Syst Integrat, Guangzhou 510006, Peoples R China
[5] Guangdong HongKong Macao Joint Lab Smart Discrete, Guangzhou 510006, Peoples R China
[6] Singapore Univ Technol & Design, Pillar Informat Syst Technol & Design, Singapore 487372, Singapore
[7] Sungkyunkwan Univ, Dept Elect & Comp Engn, Suwon 16419, South Korea
来源
20TH INTERNATIONAL WIRELESS COMMUNICATIONS & MOBILE COMPUTING CONFERENCE, IWCMC 2024 | 2024年
基金
中国国家自然科学基金; 新加坡国家研究基金会;
关键词
Generative AI (GAI); large language model; mixture of experts; network optimization;
D O I
10.1109/IWCMC61514.2024.10592370
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Optimizing various wireless user tasks poses a significant challenge for networking systems because of the expanding range of user requirements. Despite advancements in Deep Reinforcement Learning (DRL), the need for customized optimization tasks for individual users complicates developing and applying numerous DRL models, leading to substantial computation resource and energy consumption and can lead to inconsistent outcomes. To address this issue, we propose a novel approach utilizing a Mixture of Experts (MoE) framework, augmented with Large Language Models (LLMs), to analyze user objectives and constraints effectively, select specialized DRL experts, and weigh each decision from the participating experts. Specifically, we develop a gate network to oversee the expert models, allowing a collective of experts to tackle a wide array of new tasks. Furthermore, we innovatively substitute the traditional gate network with an LLM, leveraging its advanced reasoning capabilities to manage expert model selection for joint decisions. Our proposed method reduces the need to train new DRL models for each unique optimization problem, decreasing energy consumption and AI model implementation costs. The LLM-enabled MoE approach is validated through a general maze navigation task and a specific network service provider utility maximization task, demonstrating its effectiveness and practical applicability in optimizing complex networking systems.
引用
收藏
页码:531 / 536
页数:6
相关论文
共 49 条
  • [21] Collaboration between intelligent agents and large language models: A novel approach for enhancing code generation capability
    Bai, Xingyuan
    Huang, Shaobin
    Wei, Chi
    Wang, Rui
    EXPERT SYSTEMS WITH APPLICATIONS, 2025, 269
  • [22] A Large Language Model Approach to Detect Hate Speech in Political Discourse Using Multiple Language Corpora
    de Oliveira, Aillkeen Bezerra
    Baptista, Claudio de Souza
    Firmino, Anderson Almeida
    de Paiva, Anselmo Cardoso
    39TH ANNUAL ACM SYMPOSIUM ON APPLIED COMPUTING, SAC 2024, 2024, : 1461 - 1468
  • [23] Large Language Model Evaluation Criteria Framework in Healthcare: Fuzzy MCDM Approach
    Hamzeh Mohammad Alabool
    SN Computer Science, 6 (1)
  • [24] Demystify Adult Learning: A Social Network and Large Language Model Assisted Approach
    Liu, Fang
    Ding, Bosheng
    Guan, Chong
    Zhang, Wei
    Niyato, Dusit
    Tan, Justina
    2024 IEEE ANNUAL CONGRESS ON ARTIFICIAL INTELLIGENCE OF THING, AIOT 2024, 2024, : 207 - 212
  • [25] Towards Robust Neural Rankers with Large Language Model: A Contrastive Training Approach
    Pan, Ziyang
    Fan, Kangjia
    Liu, Rongyu
    Li, Daifeng
    APPLIED SCIENCES-BASEL, 2023, 13 (18):
  • [26] LLM-CloudSec: Large Language Model Empowered Automatic and Deep Vulnerability Analysis for Intelligent Clouds
    Cao, Daipeng
    Wu, Jun
    IEEE INFOCOM 2024-IEEE CONFERENCE ON COMPUTER COMMUNICATIONS WORKSHOPS, INFOCOM WKSHPS 2024, 2024,
  • [27] Redefining crowdsourced test report prioritization: An innovative approach with large language model
    Ling, Yuchen
    Yu, Shengcheng
    Fang, Chunrong
    Pan, Guobin
    Wang, Jun
    Liu, Jia
    INFORMATION AND SOFTWARE TECHNOLOGY, 2025, 179
  • [28] Evaluating Large Language Model-Assisted Emergency Triage: A Comparison of Acuity Assessments by GPT-4 and Medical Experts
    Haim, Gal Ben
    Saban, Mor
    Barash, Yiftach
    Cirulnik, David
    Shaham, Amit
    Eisenman, Ben Zion
    Burshtein, Livnat
    Mymon, Orly
    Klang, Eyal
    JOURNAL OF CLINICAL NURSING, 2024,
  • [29] Large Language Model and Digital Twins Empowered Asynchronous Federated Learning for Secure Data Sharing in Intelligent Labeling
    Sheng, Xuanzhu
    Yu, Chao
    Cui, Xiaolong
    Zhou, Yang
    MATHEMATICS, 2024, 12 (22)
  • [30] Large Language Model Enhanced Autonomous Agents for Proactive Fault-Tolerant Edge Networks
    Fang, Honglin
    Zhang, Di
    Tan, Can
    Yu, Peng
    Wang, Ying
    Li, Wenjing
    IEEE INFOCOM 2024-IEEE CONFERENCE ON COMPUTER COMMUNICATIONS WORKSHOPS, INFOCOM WKSHPS 2024, 2024,