Traffic Scenario Clustering and Load Balancing with Distilled Reinforcement Learning Policies

被引:4
|
作者
Li, Jimmy [1 ]
Wu, Di [1 ]
Xu, Yi Tian [1 ]
Li, Tianyu [1 ]
Jang, Seowoo [2 ]
Liu, Xue [1 ]
Dudek, Gregory [1 ]
机构
[1] Samsung Elect, Mississauga, ON, Canada
[2] Samsung Elect, Suwon, South Korea
关键词
load balancing; reinforcement learning; knowledge distillation;
D O I
10.1109/ICC45855.2022.9838370
中图分类号
TN [电子技术、通信技术];
学科分类号
0809 ;
摘要
Due to the rapid increase in wireless communication traffic in recent years, load balancing is becoming increasingly important for ensuring the quality of service. However, variations in traffic patterns near different serving base stations make this task challenging. On one hand, crafting a single control policy that performs well across all base station sectors is often difficult. On the other hand, maintaining separate controllers for every sector introduces overhead, and leads to redundancy if some of the sectors experience similar traffic patterns. In this paper, we propose to construct a concise set of controllers that cover a wide range of traffic scenarios, allowing the operator to select a suitable controller for each sector based on local traffic conditions. To construct these controllers, we present a method that clusters similar scenarios and learns a general control policy for each cluster. We use deep reinforcement learning (RL) to first train separate control policies on diverse traffic scenarios, and then incrementally merge together similar RL policies via knowledge distillation. Experimental results show that our concise policy set reduces redundancy with very minor performance degradation compared to policies trained separately on each traffic scenario. Our method also outperforms hand-crafted control parameters, joint learning on all tasks, and two popular clustering methods.
引用
收藏
页码:1536 / 1541
页数:6
相关论文
共 50 条
  • [31] A reinforcement learning-based load balancing algorithm for fog computing
    Niloofar Tahmasebi-Pouya
    Mehdi Agha Sarram
    Seyedakbar Mostafavi
    Telecommunication Systems, 2023, 84 : 321 - 339
  • [32] Reinforcement Learning to Improve Resource Scheduling and Load Balancing in Cloud Computing
    Kaveri P.R.
    Lahande P.
    SN Computer Science, 4 (2)
  • [33] Reinforcement Learning in the Load Balancing Problem for the iFDAQ of the COMPASS Experiment at CERN
    Subrt, Ondrej
    Bodlak, Martin
    Jandek, Matous
    Jary, Vladimir
    Kveton, Antonin
    Novy, Josef
    Virius, Miroslav
    Zemko, Martin
    ICAART: PROCEEDINGS OF THE 12TH INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE, VOL 2, 2020, : 734 - 741
  • [34] A Load Balancing Scheme for Gaming Server applying Reinforcement Learning in IoT
    Kim, Hye-Young
    Kim, Jinsul
    COMPUTER SCIENCE AND INFORMATION SYSTEMS, 2020, 17 (03) : 891 - 906
  • [35] Optimizing SDN Controller Load Balancing Using Online Reinforcement Learning
    Kumari, Abha
    Roy, Arghyadip
    Sairam, Ashok Singh
    IEEE ACCESS, 2024, 12 : 131591 - 131604
  • [36] Deep Reinforcement Learning for Load Balancing of Edge Servers in IoV (2022)
    Li, Pu
    Xie, Wenxuan
    Yuan, Ying
    Chen, Chen
    Wan, Shaohua
    MOBILE NETWORKS & APPLICATIONS, 2022, 27 (04): : 1475 - 1475
  • [37] A reinforcement learning-based load balancing algorithm for fog computing
    Tahmasebi-Pouya, Niloofar
    Sarram, Mehdi Agha
    Mostafavi, Seyedakbar
    TELECOMMUNICATION SYSTEMS, 2023, 84 (03) : 321 - 339
  • [38] Load Balancing of Hybrid LiFi WiFi Networks Using Reinforcement learning
    Ahmad, Rizwana
    Soltani, Mohammad Dehghani
    Safari, Majid
    Srivastava, Anand
    2020 IEEE 31ST ANNUAL INTERNATIONAL SYMPOSIUM ON PERSONAL, INDOOR AND MOBILE RADIO COMMUNICATIONS (IEEE PIMRC), 2020,
  • [39] Traffic load monitoring and load balancing for the Internet
    Bing Chen
    Shahram Latifi
    Cluster Computing, 2000, 3 (2) : 139 - 150
  • [40] Enhancing fog load balancing through lifelong transfer learning of reinforcement learning agents
    Ebrahim, Maad
    Hafid, Abdelhakim
    Abid, Mohamed Riduan
    COMPUTER COMMUNICATIONS, 2025, 231