A Lightweight Reinforcement Learning Based Packet Routing Method Using Online Sequential Learning

被引:1
|
作者
Nemoto, Kenji [1 ]
Matsutani, Hiroki [1 ]
机构
[1] Keio Univ, Grad Sch Sci & Technol, Yokohama 2238522, Japan
关键词
reinforcement learning; packet routing; neural networks; OS-ELM;
D O I
10.1587/transinf.2022EDP7231
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Existing simple routing protocols (e.g., OSPF, RIP) have some disadvantages of being inflexible and prone to congestion due to the concentration of packets on particular routers. To address these issues, packet routing methods using machine learning have been proposed recently. Compared to these algorithms, machine learning based methods can choose a routing path intelligently by learning efficient routes. However, machine learning based methods have a disadvantage of training time overhead. We thus focus on a lightweight machine learning algorithm, OS-ELM (Online Sequential Extreme Learning Machine), to reduce the training time. Although previous work on reinforcement learning using OS-ELM exists, it has a problem of low learning accuracy. In this paper, we propose OS-ELM QN (Q-Network) with a prioritized experience replay buffer to improve the learning performance. It is compared to a deep reinforcement learning based packet routing method using a network simulator. Experimental results show that introducing the experience replay buffer improves the learning performance. OS-ELM QN achieves a 2.33 times speedup than a DQN (Deep Q-Network) in terms of learning speed. Regarding the packet transfer latency, OS-ELM QN is comparable or slightly inferior to the DQN while they are better than OSPF in most cases since they can distribute congestions.
引用
收藏
页码:1796 / 1807
页数:12
相关论文
共 50 条
  • [41] Online Observer-Based Inverse Reinforcement Learning
    Self, Ryan
    Coleman, Kevin
    Bai, He
    Kamalapurkar, Rushikesh
    IEEE CONTROL SYSTEMS LETTERS, 2021, 5 (06): : 1922 - 1927
  • [42] Sequential and Dynamic constraint Contrastive Learning for Reinforcement Learning
    Shen, Weijie
    Yuan, Lei
    Huang, Junfu
    Gao, Songyi
    Huang, Yuyang
    Yu, Yang
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [43] Routing in quantum communication networks using reinforcement machine learning
    Roik, Jan
    Bartkiewicz, Karol
    Cernoch, Antonin
    Lemr, Karel
    QUANTUM INFORMATION PROCESSING, 2024, 23 (03)
  • [44] Reinforcement Learning Based on Contextual Bandits for Personalized Online Learning Recommendation Systems
    Wacharawan Intayoad
    Chayapol Kamyod
    Punnarumol Temdee
    Wireless Personal Communications, 2020, 115 : 2917 - 2932
  • [45] Online Reinforcement Learning-Based Strategy Learning in Iterated Prisoners Dilemma
    Xing, Xiaoyu
    Xia, Haoxiang
    IEEE TRANSACTIONS ON COMPUTATIONAL SOCIAL SYSTEMS, 2025,
  • [46] Routing in quantum communication networks using reinforcement machine learning
    Jan Roik
    Karol Bartkiewicz
    Antonín Černoch
    Karel Lemr
    Quantum Information Processing, 23
  • [47] Decentralized Covert Routing in Heterogeneous Networks Using Reinforcement Learning
    Kong, Justin
    Moore, Terrence J.
    Dagefu, Fikadu T.
    IEEE COMMUNICATIONS LETTERS, 2024, 28 (11) : 2683 - 2687
  • [48] Using Reinforcement Learning to Perform Qubit Routing in Quantum Compilers
    Pozzi, Matteo G.
    Herbert, Steven J.
    Sengupta, Akash
    Mullins, Robert D.
    ACM TRANSACTIONS ON QUANTUM COMPUTING, 2022, 3 (02):
  • [49] Reinforcement Learning Based on Contextual Bandits for Personalized Online Learning Recommendation Systems
    Intayoad, Wacharawan
    Kamyod, Chayapol
    Temdee, Punnarumol
    WIRELESS PERSONAL COMMUNICATIONS, 2020, 115 (04) : 2917 - 2932
  • [50] Fault-Tolerant Routing Algorithm for Mesh based NoC using Reinforcement Learning
    Samala, Jagadheesh
    Takawale, Harshvardhan
    Chokhani, Yash
    Bhanu, P. Veda
    Soumya, J.
    2020 24TH INTERNATIONAL SYMPOSIUM ON VLSI DESIGN AND TEST (VDAT), 2020,