Accelerating decentralized reinforcement learning of complex individual behaviors

被引:2
|
作者
Leottau, David L. [1 ]
Lobos-Tsunekawa, Kenzo [1 ]
Jaramillo, Francisco [1 ]
Ruiz-del-Solar, Javier [1 ]
机构
[1] Univ Chile, Adv Min Technol Ctr, Dept Elect Engn, Ave Tupper 2007, Santiago, Chile
关键词
Decentralized reinforcement learning; Multi-agent systems; Distributed control; Autonomous robots; Knowledge transfer; Distributed artificial intelligence;
D O I
10.1016/j.engappai.2019.06.019
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Many Reinforcement Learning (RL) real-world applications have multi-dimensional action spaces which suffer from the combinatorial explosion of complexity. Then, it may turn infeasible to implement Centralized RL (CRL) systems due to the exponential increasing of dimensionality in both the state space and the action space, and the large number of training trials. In order to address this, this paper proposes to deal with these issues by using Decentralized Reinforcement Learning (DRL) to alleviate the effects of the curse of dimensionality on the action space, and by transferring knowledge to reduce the training episodes so that asymptotic converge can be achieved. Three DRL schemes are compared: DRL with independent learners and no prior-coordination (DRLInd); DRL accelerated-coordinated by using the Control Sharing (DRL+CoSh) Knowledge Transfer approach; and a proposed DRL scheme using the CoSh-based variant Nearby Action Sharing to include a measure of the uncertainty into the CoSh procedure (DRL+NeASh). These three schemes are analyzed through an extensive experimental study and validated through two complex real-world problems, namely the inwalk-kicking and the ball-dribbling behaviors, both performed with humanoid biped robots. Obtained results show (empirically): (i) the effectiveness of DRL systems which even without prior-coordination are able to achieve asymptotic convergence throughout indirect coordination; (ii) that by using the proposed knowledge transfer methods, it is possible to reduce the training episodes and to coordinate the DRL process; and (iii) obtained learning times are between 36% and 62% faster than the DRL-Ind schemes in the case studies.
引用
收藏
页码:243 / 253
页数:11
相关论文
共 50 条
  • [31] Designing Decentralized Controllers for Distributed-Air-Jet MEMS-Based Micromanipulators by Reinforcement Learning
    Laëtitia Matignon
    Guillaume J. Laurent
    Nadine Le Fort-Piat
    Yves-André Chapuis
    Journal of Intelligent & Robotic Systems, 2010, 59 : 145 - 166
  • [32] Distributed Reinforcement Learning for Decentralized Linear Quadratic Control: A Derivative-Free Policy Optimization Approach
    Li, Yingying
    Tang, Yujie
    Zhang, Runyu
    Li, Na
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2022, 67 (12) : 6429 - 6444
  • [33] Designing Decentralized Controllers for Distributed-Air-Jet MEMS-Based Micromanipulators by Reinforcement Learning
    Matignon, Laetitia
    Laurent, Guillaume J.
    Le Fort-Piat, Nadine
    Chapuis, Yves-Andr
    JOURNAL OF INTELLIGENT & ROBOTIC SYSTEMS, 2010, 59 (02) : 145 - 166
  • [34] Cooperative Deep Reinforcement Learning Policies for Autonomous Navigation in Complex Environments
    Tran, Van Manh
    Kim, Gon-Woo
    IEEE ACCESS, 2024, 12 : 101053 - 101065
  • [35] A comprehensive survey of multiagent reinforcement learning
    Busoniu, Lucian
    Babuska, Robert
    De Schutter, Bart
    IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART C-APPLICATIONS AND REVIEWS, 2008, 38 (02): : 156 - 172
  • [36] Online Multi-Agent Reinforcement Learning for Decentralized Inverter-Based Volt-VAR Control
    Liu, Haotian
    Wu, Wenchuan
    IEEE TRANSACTIONS ON SMART GRID, 2021, 12 (04) : 2980 - 2990
  • [37] Finite-Sample Analysis For Decentralized Cooperative Multi-Agent Reinforcement Learning From Batch Data
    Zhang, Kaiqing
    Yang, Zhuoran
    Liu, Han
    Zhang, Tong
    Basar, Tamer
    IFAC PAPERSONLINE, 2020, 53 (02): : 1049 - 1056
  • [38] Detect to Focus: Latent-Space Autofocusing System With Decentralized Hierarchical Multi-Agent Reinforcement Learning
    Anikina, Anna
    Rogov, Oleg Y.
    Dylov, Dmitry V.
    IEEE ACCESS, 2023, 11 : 85214 - 85223
  • [39] Learning to communicate in a decentralized environment
    Claudia V. Goldman
    Martin Allen
    Shlomo Zilberstein
    Autonomous Agents and Multi-Agent Systems, 2007, 15 : 47 - 90
  • [40] Learning to communicate in a decentralized environment
    Goldman, Claudia V.
    Allen, Martin
    Zilberstein, Shlomo
    AUTONOMOUS AGENTS AND MULTI-AGENT SYSTEMS, 2007, 15 (01) : 47 - 90