Coordinated behavior of cooperative agents using deep reinforcement learning

被引:11
|
作者
Diallo, Elhadji Amadou Oury [1 ]
Sugiyama, Ayumi [1 ]
Sugawara, Toshiharu [1 ]
机构
[1] Waseda Univ, Dept Comp Sci & Commun Engn, Shinjuku Ku, 3-4-1 Okubo, Tokyo 1698555, Japan
关键词
Deep reinforcement learning; Multi-agent systems; Cooperation; Coordination; LINEAR MULTIAGENT SYSTEMS; INTELLIGENCE; TAXONOMY;
D O I
10.1016/j.neucom.2018.08.094
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this work, we focus on an environment where multiple agents with complementary capabilities cooperate to generate non-conflicting joint actions that achieve a specific target. The central problem addressed is how several agents can collectively learn to coordinate their actions such that they complete a given task together without conflicts. However, sequential decision-making under uncertainty is one of the most challenging issues for intelligent cooperative systems. To address this, we propose a multi-agent concurrent framework where agents learn coordinated behaviors in order to divide their areas of responsibility. The proposed framework is an extension of some recent deep reinforcement learning algorithms such as DQN, double DQN, and dueling network architectures. Then, we investigate how the learned behaviors change according to the dynamics of the environment, reward scheme, and network structures. Next, we show how agents behave and choose their actions such that the resulting joint actions are optimal. We finally show that our method can lead to stable solutions in our specific environment. (C) 2019 Elsevier B.V. All rights reserved.
引用
收藏
页码:230 / 240
页数:11
相关论文
共 50 条
  • [21] Emergence of Cooperative Automated Driving Control at Roundabouts Using Deep Reinforcement Learning
    Nakaya, Reo
    Harada, Tomohiro
    Miura, Yukiya
    Hattori, Kiyohiko
    Matsuoka, Johei
    2023 62ND ANNUAL CONFERENCE OF THE SOCIETY OF INSTRUMENT AND CONTROL ENGINEERS, SICE, 2023, : 97 - 102
  • [22] Goal Modelling for Deep Reinforcement Learning Agents
    Leung, Jonathan
    Shen, Zhiqi
    Zeng, Zhiwei
    Miao, Chunyan
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, 2021, 12975 : 271 - 286
  • [23] Behavior fusion for deep reinforcement learning
    Shi, Haobin
    Xu, Meng
    Hwang, Kao-Shing
    Cai, Bo-Yin
    ISA TRANSACTIONS, 2020, 98 : 434 - 444
  • [24] Coordinated Energy and Reserve Sharing of Isolated Microgrid Cluster using Deep Reinforcement Learning
    Qazi, Hasan Saeed
    Liu, Nian
    Wang, Tong
    2020 5TH ASIA CONFERENCE ON POWER AND ELECTRICAL ENGINEERING (ACPEE 2020), 2020, : 81 - 86
  • [25] Decentralized Scheduling for Cooperative Localization With Deep Reinforcement Learning
    Peng, Bile
    Seco-Granados, Gonzalo
    Steinmetz, Erik
    Frohle, Markus
    Wymeersch, Henk
    IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2019, 68 (05) : 4295 - 4305
  • [26] Training Cooperative Agents for Multi-Agent Reinforcement Learning
    Bhalla, Sushrut
    Subramanian, Sriram G.
    Crowley, Mark
    AAMAS '19: PROCEEDINGS OF THE 18TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS, 2019, : 1826 - 1828
  • [27] Online Surveillance of IoT Agents in Smart Cities Using Deep Reinforcement Learning
    Alenezi, Ahmad
    INTERNATIONAL JOURNAL OF INTELLIGENT INFORMATION TECHNOLOGIES, 2024, 20 (01)
  • [28] FLARE: Fingerprinting Deep Reinforcement Learning Agents using Universal Adversarial Masks
    Tekgul, Buse G. A.
    Asokan, N.
    39TH ANNUAL COMPUTER SECURITY APPLICATIONS CONFERENCE, ACSAC 2023, 2023, : 492 - 505
  • [29] Optimization for Reinforcement Learning: From a single agent to cooperative agents
    Lee, Donghwan
    He, Niao
    Kamalaruban, Parameswaran
    Cevher, Volkan
    IEEE SIGNAL PROCESSING MAGAZINE, 2020, 37 (03) : 123 - 135
  • [30] Microscopic Traffic Simulation by Cooperative Multi-agent Deep Reinforcement Learning
    Bacchiani, Giulio
    Molinari, Daniele
    Patander, Marco
    AAMAS '19: PROCEEDINGS OF THE 18TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS, 2019, : 1547 - 1555