Evaluating cooperative-competitive dynamics with deep Q-learning

被引:2
|
作者
Kopacz, Aniko [1 ]
Csato, Lehel [1 ]
Chira, Camelia [1 ]
机构
[1] Babes Bolyai Univ, Fac Math & Comp Sci, 1 Mihail Kogalniceanu Str, RO-400084 Cluj Napoca, Romania
关键词
Multi -agent systems; Reinforcement learning; Deep Q -learning;
D O I
10.1016/j.neucom.2023.126507
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We model cooperative-competitive social group dynamics with multi-agent environments, specialized in cases with a large number of agents from only a few distinct types. The multi-agent optimization problems are addressed in turn with multi-agent reinforcement learning algorithms to obtain flexible and robust solutions. We analyze the effectiveness of centralized and decentralized algorithms using three variants of deep Q-networks on these cooperative-competitive environments: first, we use the decentralized training independent learning with deep Q-networks, secondly the centralized monotonic value factorizations for deep learning, and lastly the multi-agent variational exploration. We test the algorithms in simulated predator-prey multi-agent environments in two distinct environments: the adversary pursuit and simple tag. The experiments highlight the performance of the different deep Q-learning methods, and we conclude that decentralized training of deep Q-networks accumulates higher episode rewards during training and evaluation in comparison with the selected centralized learning approaches.& COPY; 2023 Elsevier B.V. All rights reserved.
引用
收藏
页数:8
相关论文
共 50 条
  • [21] Deep Q-Learning with Prioritized Sampling
    Zhai, Jianwei
    Liu, Quan
    Zhang, Zongzhang
    Zhong, Shan
    Zhu, Haijun
    Zhang, Peng
    Sun, Cijia
    NEURAL INFORMATION PROCESSING, ICONIP 2016, PT I, 2016, 9947 : 13 - 22
  • [22] Cooperative Deep Q-Learning With Q-Value Transfer for Multi-Intersection Signal Control
    Ge, Hongwei
    Song, Yumei
    Wu, Chunguo
    Ren, Jiankang
    Tan, Guozhen
    IEEE ACCESS, 2019, 7 : 40797 - 40809
  • [23] Mixed Cooperative-Competitive Communication Using Multi-agent Reinforcement Learning
    Vanneste, Astrid
    Van Wijnsberghe, Wesley
    Vanneste, Simon
    Mets, Kevin
    Mercelis, Siegfried
    Latre, Steven
    Hellinckx, Peter
    ADVANCES ON P2P, PARALLEL, GRID, CLOUD AND INTERNET COMPUTING, 3PGCIC-2021, 2022, 343 : 197 - 206
  • [24] Cooperative-Competitive Decision-Making in Resource Management: A Reinforcement Learning Perspective
    Isakov, Artem
    Peregorodiev, Danil
    Brunko, Pavel
    Tomilov, Ivan
    Gusarova, Natalia
    Vatian, Alexandra
    INTELLIGENT DATA ENGINEERING AND AUTOMATED LEARNING - IDEAL 2024, PT I, 2025, 15346 : 375 - 386
  • [25] Evaluating Action Durations for Adaptive Traffic Signal Control Based On Deep Q-Learning
    Celtek, Seyit Alperen
    Durdu, Akif
    Ali, Muzamil Eltejani Mohammed
    INTERNATIONAL JOURNAL OF INTELLIGENT TRANSPORTATION SYSTEMS RESEARCH, 2021, 19 (03) : 557 - 571
  • [26] Evaluating Action Durations for Adaptive Traffic Signal Control Based On Deep Q-Learning
    Seyit Alperen CELTEK
    Akif DURDU
    Muzamil Eltejani Mohammed ALI
    International Journal of Intelligent Transportation Systems Research, 2021, 19 : 557 - 571
  • [27] A Cooperative-Competitive Strategy for Autonomous Multidrone Racing
    Di, Jian
    Chen, Shaofeng
    Li, Pengfei
    Wang, Xinghu
    Ji, Haibo
    Kang, Yu
    IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2024, 71 (07) : 7488 - 7497
  • [28] Expected Lenient Q-learning: a fast variant of the Lenient Q-learning algorithm for cooperative stochastic Markov games
    Amhraoui, Elmehdi
    Masrour, Tawfik
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2024, 15 (07) : 2781 - 2797
  • [29] Multi-goal Q-learning of cooperative teams
    Li, Jing
    Sheng, Zhaohan
    Ng, KwanChew
    EXPERT SYSTEMS WITH APPLICATIONS, 2011, 38 (03) : 1565 - 1574
  • [30] Distributed lazy Q-learning for cooperative mobile robots
    Touzet, Claude F.
    International Journal of Advanced Robotic Systems, 2004, 1 (01) : 5 - 13