Evaluating cooperative-competitive dynamics with deep Q-learning

被引:2
|
作者
Kopacz, Aniko [1 ]
Csato, Lehel [1 ]
Chira, Camelia [1 ]
机构
[1] Babes Bolyai Univ, Fac Math & Comp Sci, 1 Mihail Kogalniceanu Str, RO-400084 Cluj Napoca, Romania
关键词
Multi -agent systems; Reinforcement learning; Deep Q -learning;
D O I
10.1016/j.neucom.2023.126507
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We model cooperative-competitive social group dynamics with multi-agent environments, specialized in cases with a large number of agents from only a few distinct types. The multi-agent optimization problems are addressed in turn with multi-agent reinforcement learning algorithms to obtain flexible and robust solutions. We analyze the effectiveness of centralized and decentralized algorithms using three variants of deep Q-networks on these cooperative-competitive environments: first, we use the decentralized training independent learning with deep Q-networks, secondly the centralized monotonic value factorizations for deep learning, and lastly the multi-agent variational exploration. We test the algorithms in simulated predator-prey multi-agent environments in two distinct environments: the adversary pursuit and simple tag. The experiments highlight the performance of the different deep Q-learning methods, and we conclude that decentralized training of deep Q-networks accumulates higher episode rewards during training and evaluation in comparison with the selected centralized learning approaches.& COPY; 2023 Elsevier B.V. All rights reserved.
引用
收藏
页数:8
相关论文
共 50 条
  • [1] Deep Reinforcement Learning: From Q-Learning to Deep Q-Learning
    Tan, Fuxiao
    Yan, Pengfei
    Guan, Xinping
    NEURAL INFORMATION PROCESSING (ICONIP 2017), PT IV, 2017, 10637 : 475 - 483
  • [2] A switched model for mixed cooperative-competitive social dynamics
    Blanchini, Franco
    Casagrande, Daniele
    Giordano, Giulia
    Viaro, Umberto
    2019 18TH EUROPEAN CONTROL CONFERENCE (ECC), 2019, : 3514 - 3519
  • [3] Q-learning in a competitive supply chain
    van Tongeren, Tim
    Kaymak, Uzay
    Naso, David
    van Asperen, Eelco
    2007 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN AND CYBERNETICS, VOLS 1-8, 2007, : 3313 - +
  • [4] Cooperative pursuit with multiple pursuers based on Deep Minimax Q-learning
    Ji, Mengda
    Xu, Genjiu
    Duan, Zekun
    Wang, Liying
    Li, Zesheng
    Ge, Jianjun
    Li, Mingqiang
    AEROSPACE SCIENCE AND TECHNOLOGY, 2024, 146
  • [5] Cooperative Deep Q-Learning Framework for Environments Providing Image Feedback
    Raghavan, Krishnan
    Narayanan, Vignesh
    Jagannathan, Sarangapani
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (07) : 9267 - 9276
  • [6] Cooperative Q-Learning Based on Learning Automata
    Yang, Mao
    Tian, Yantao
    Qi, Xinyue
    2009 IEEE INTERNATIONAL CONFERENCE ON AUTOMATION AND LOGISTICS ( ICAL 2009), VOLS 1-3, 2009, : 1972 - 1977
  • [7] EFFECTS OF COMMUNICATION IN COOPERATIVE Q-LEARNING
    Darbyshire, Paul
    Wang, Dianhui
    INTERNATIONAL JOURNAL OF INNOVATIVE COMPUTING INFORMATION AND CONTROL, 2010, 6 (05): : 2113 - 2126
  • [8] Constrained Deep Q-Learning Gradually Approaching Ordinary Q-Learning
    Ohnishi, Shota
    Uchibe, Eiji
    Yamaguchi, Yotaro
    Nakanishi, Kosuke
    Yasui, Yuji
    Ishii, Shin
    FRONTIERS IN NEUROROBOTICS, 2019, 13
  • [9] Expertness based cooperative Q-learning
    Ahmadabadi, MN
    Asadpour, M
    IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART B-CYBERNETICS, 2002, 32 (01): : 66 - 76
  • [10] Distributed Cognitive Learning Strategy for Cooperative-Competitive Multiagent Systems
    Liu, Yan-Jun
    Zhang, Sai
    Tang, Li
    IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2024, 16 (01) : 20 - 30