Multi-agent Reinforcement Learning in a Large Scale Environment via Supervisory Network and Curriculum Learning

被引:0
|
作者
Do, Seungwon [1 ]
Lee, Changeun [1 ]
机构
[1] Elect & Telecommun Res Inst ETRI, Def ICT Convers Res Sect, Daejeon 34129, South Korea
来源
2021 21ST INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND SYSTEMS (ICCAS 2021) | 2021年
关键词
Deep reinforcement learning; action planning; task assignment; and curriculum learning;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Multi-agent reinforcement learning is essential for learning optimal policy for collaboration and competition environments. However, as the action space of the agent increases, the number of state-action pairs which have to be explored increases exponentially. As a result, increasing search space causes difficulty to converge the learning. To solve this problem, we propose a supervisory network. To achieve the global goal, the supervisory network creates a sub-goal and assigns the goals to the agents so that the agents can effectively learn the optimal policy with a small action space. In addition, we adapt the curriculum learning method to learn a large-scale environment. As a consequence, the agent can explore the environment in which the complexity increases gradually. Although a baseline network was learned in the same environment to compare with our model, the baseline fails to learn an optimal policy while our model successes to learn in the large-scale environment.
引用
收藏
页码:207 / 210
页数:4
相关论文
共 50 条
  • [1] Pacesetter Learning for Large Scale Cooperative Multi-Agent Reinforcement Learning
    Zhou, Pingqi
    Li, Chao
    Qiu, Mengwei
    Liu, Jun
    Ma, Chennan
    Yan, Ming
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2023, PT VI, 2023, 14259 : 115 - 126
  • [2] Concentration Network for Reinforcement Learning of Large-Scale Multi-Agent Systems
    Fu, Qingxu
    Qiu, Tenghai
    Yi, Jianqiang
    Pu, Zhiqiang
    Wu, Shiguang
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 9341 - 9349
  • [3] Cooperative Multi-Agent Reinforcement Learning in a Large Stationary Environment
    Zemzem, Wiem
    Tagina, Moncef
    2017 16TH IEEE/ACIS INTERNATIONAL CONFERENCE ON COMPUTER AND INFORMATION SCIENCE (ICIS 2017), 2017, : 365 - 371
  • [4] CityFlow: A Multi-Agent Reinforcement Learning Environment for Large Scale City Traffic Scenario
    Zhang, Huichu
    Feng, Siyuan
    Liu, Chang
    Ding, Yaoyao
    Zhu, Yichen
    Zhou, Zihan
    Zhang, Weinan
    Yu, Yong
    Jin, Haiming
    Li, Zhenhui
    WEB CONFERENCE 2019: PROCEEDINGS OF THE WORLD WIDE WEB CONFERENCE (WWW 2019), 2019, : 3620 - 3624
  • [5] Large-Scale Machine Learning Cluster Scheduling via Multi-Agent Graph Reinforcement Learning
    Zhao, Xiaoyang
    Wu, Chuan
    IEEE TRANSACTIONS ON NETWORK AND SERVICE MANAGEMENT, 2022, 19 (04): : 4962 - 4974
  • [6] Network Maintenance Planning Via Multi-Agent Reinforcement Learning
    Thomas, Jonathan
    Hernandez, Marco Perez
    Parlikad, Ajith Kumar
    Piechocki, Robert
    2021 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2021, : 2289 - 2295
  • [7] Learning adversarial policy in multiple scenes environment via multi-agent reinforcement learning
    Li, Yang
    Wang, Xinzhi
    Wang, Wei
    Zhang, Zhenyu
    Wang, Jianshu
    Luo, Xiangfeng
    Xie, Shaorong
    CONNECTION SCIENCE, 2021, 33 (03) : 407 - 426
  • [8] A Curriculum Framework for Autonomous Network Defense using Multi-agent Reinforcement Learning
    Campbell, Robert G.
    Eirinaki, Magdalini
    Park, Younghee
    2023 SILICON VALLEY CYBERSECURITY CONFERENCE, SVCC, 2023,
  • [9] Multi-agent Reinforcement Learning in Network Management
    Bagnasco, Ricardo
    Serrat, Joan
    SCALABILITY OF NETWORKS AND SERVICES, PROCEEDINGS, 2009, 5637 : 199 - 202
  • [10] SCM network with multi-agent reinforcement learning
    Zhao, Gang
    Sun, Ruoying
    FIFTH WUHAN INTERNATIONAL CONFERENCE ON E-BUSINESS, VOLS 1-3, 2006, : 1294 - 1300