Evaluating cooperative-competitive dynamics with deep Q-learning

被引:2
|
作者
Kopacz, Aniko [1 ]
Csato, Lehel [1 ]
Chira, Camelia [1 ]
机构
[1] Babes Bolyai Univ, Fac Math & Comp Sci, 1 Mihail Kogalniceanu Str, RO-400084 Cluj Napoca, Romania
关键词
Multi -agent systems; Reinforcement learning; Deep Q -learning;
D O I
10.1016/j.neucom.2023.126507
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We model cooperative-competitive social group dynamics with multi-agent environments, specialized in cases with a large number of agents from only a few distinct types. The multi-agent optimization problems are addressed in turn with multi-agent reinforcement learning algorithms to obtain flexible and robust solutions. We analyze the effectiveness of centralized and decentralized algorithms using three variants of deep Q-networks on these cooperative-competitive environments: first, we use the decentralized training independent learning with deep Q-networks, secondly the centralized monotonic value factorizations for deep learning, and lastly the multi-agent variational exploration. We test the algorithms in simulated predator-prey multi-agent environments in two distinct environments: the adversary pursuit and simple tag. The experiments highlight the performance of the different deep Q-learning methods, and we conclude that decentralized training of deep Q-networks accumulates higher episode rewards during training and evaluation in comparison with the selected centralized learning approaches.& COPY; 2023 Elsevier B.V. All rights reserved.
引用
收藏
页数:8
相关论文
共 50 条
  • [31] Reinforcement distribution in a team of cooperative Q-learning agents
    Abbasi, Zahra
    Abbasi, Mohammad Ali
    PROCEEDINGS OF NINTH ACIS INTERNATIONAL CONFERENCE ON SOFTWARE ENGINEERING, ARTIFICIAL INTELLIGENCE, NETWORKING AND PARALLEL/DISTRIBUTED COMPUTING, 2008, : 154 - +
  • [32] Q-Learning for Content Placement in Wireless Cooperative Caching
    Yang, Zhong
    Liu, Yuanwei
    Chen, Yue
    2018 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2018,
  • [33] Deep Q-Learning with Phased Experience Cooperation
    Wang, Hongbo
    Zeng, Fanbing
    Tu, Xuyan
    COMPUTER SUPPORTED COOPERATIVE WORK AND SOCIAL COMPUTING, CHINESECSCW 2019, 2019, 1042 : 752 - 765
  • [34] Deep Q-learning: A robust control approach
    Varga, Balazs
    Kulcsar, Balazs
    Chehreghani, Morteza Haghir
    INTERNATIONAL JOURNAL OF ROBUST AND NONLINEAR CONTROL, 2023, 33 (01) : 526 - 544
  • [35] Stochastic Variance Reduction for Deep Q-learning
    Zhao, Wei-Ye
    Peng, Jian
    AAMAS '19: PROCEEDINGS OF THE 18TH INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS AND MULTIAGENT SYSTEMS, 2019, : 2318 - 2320
  • [36] Deep Surrogate Q-Learning for Autonomous Driving
    Kalweit, Maria
    Kalweit, Gabriel
    Werling, Moritz
    Boedecker, Joschka
    2022 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2022), 2022, : 1578 - 1584
  • [37] Trading ETFs with Deep Q-Learning Algorithm
    Hong, Shao-Yan
    Liu, Chien-Hung
    Chen, Woei-Kae
    You, Shingchern D.
    2020 IEEE INTERNATIONAL CONFERENCE ON CONSUMER ELECTRONICS - TAIWAN (ICCE-TAIWAN), 2020,
  • [38] Deep Q-Learning for Aggregator Price Design
    Pigott, Aisling
    Baker, Kyri
    Mosiman, Cory
    2021 IEEE POWER & ENERGY SOCIETY GENERAL MEETING (PESGM), 2021,
  • [39] Diagnosing Bottlenecks in Deep Q-learning Algorithms
    Fu, Justin
    Kumar, Aviral
    Soh, Matthew
    Levine, Sergey
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [40] Two-Step Deep Reinforcement Q-Learning based Relay Selection in Cooperative WPCNs
    Tolebi, Gulnur
    Tsiftsis, Theodoros A.
    Nauryzbayev, Galymzhan
    2023 INTERNATIONAL BALKAN CONFERENCE ON COMMUNICATIONS AND NETWORKING, BALKANCOM, 2023,