Effect of Virtual Work Braking on Distributed Multi-Robot Reinforcement Learning

被引:0
|
作者
Kawano, Hiroshi [1 ]
机构
[1] NTT Corp, NTT Commun Sci Labs, Kanagawa, Japan
来源
2013 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC 2013) | 2013年
关键词
multi-agent learning; reinforcement learning; robotics; embodiment; distributed system;
D O I
10.1109/SMC.2013.341
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Multi-agent reinforcement learning (MARL) is one of the most promising methods for solving the problem of multi-robot control. One approach for MARL is cooperative Q-learning (CoQ), which uses learning state space containing states and actions of all agents. Inspite of the mathematical foundation for learning convergence, CoQ often suffers from a state space explosion caused by the increase in the number of agents. Another approach to MARL is distributed Q-learning (DiQ), in which each agent uses learning state space not containing the states and actions of other agents. The state space for DiQ can easily be kept compact. Therefore, DiQ seems suitable for solving multi-robot control problems. However, there is no mathematical guarantee for learning convergence in DiQ and it is difficult to apply DiQ to a multi-robot control problem in which definite appointments among working robots must be considered for accomplishing a mission. To solve these problems in applying DiQ for multi-robot control, we treat the work operated by robots as a new agent that regulates robots' motion. We assume that the work has braking ability for its motion. The work stops its motion when the robot attempts to push the work in an inappropriate direction. The policy for the work braking is obtained via dynamic programming of a Markov decision process by using a map of the environment and the work's geometry. By virtue of this, DiQ without joint state space shows convergence. Simulation results also show the high performance of the proposed method in learning speed.
引用
收藏
页码:1987 / 1994
页数:8
相关论文
共 50 条
  • [1] Distributed Reinforcement Learning for Coordinate Multi-Robot Foraging
    Hongliang Guo
    Yan Meng
    Journal of Intelligent & Robotic Systems, 2010, 60 : 531 - 551
  • [2] Distributed Reinforcement Learning for Coordinate Multi-Robot Foraging
    Guo, Hongliang
    Meng, Yan
    JOURNAL OF INTELLIGENT & ROBOTIC SYSTEMS, 2010, 60 (3-4) : 531 - 551
  • [3] Distributed safe reinforcement learning for multi-robot motion planning
    Lu, Yang
    Guo, Yaohua
    Zhao, Guoxiang
    Zhu, Minghui
    2021 29TH MEDITERRANEAN CONFERENCE ON CONTROL AND AUTOMATION (MED), 2021, : 1209 - 1214
  • [4] Reinforcement learning in the multi-robot domain
    Mataric, MJ
    AUTONOMOUS ROBOTS, 1997, 4 (01) : 73 - 83
  • [5] Reinforcement Learning in the Multi-Robot Domain
    Maja J. Matarić
    Autonomous Robots, 1997, 4 : 73 - 83
  • [6] Distributed multi-agent deep reinforcement learning for cooperative multi-robot pursuit
    Yu, Chao
    Dong, Yinzhao
    Li, Yangning
    Chen, Yatong
    JOURNAL OF ENGINEERING-JOE, 2020, 2020 (13): : 499 - 504
  • [7] Cooperative Multi-Robot Hierarchical Reinforcement Learning
    Setyawan, Gembong Edhi
    Hartono, Pitoyo
    Sawada, Hideyuki
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2022, 13 (09) : 35 - 44
  • [8] Multi-robot Teleoperation Based On Distributed Virtual Environment
    Gao, Sheng
    Chen, Dongdong
    FRONTIERS OF MANUFACTURING SCIENCE AND MEASURING TECHNOLOGY III, PTS 1-3, 2013, 401 : 1923 - 1926
  • [9] LEMURS: Learning Distributed Multi-Robot Interactions
    Sebastian, Eduardo
    Duong, Thai
    Atanasov, Nikolay
    Montijano, Eduardo
    Sagues, Carlos
    2023 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2023), 2023, : 7713 - 7719