Evaluating cooperative-competitive dynamics with deep Q-learning

被引:2
|
作者
Kopacz, Aniko [1 ]
Csato, Lehel [1 ]
Chira, Camelia [1 ]
机构
[1] Babes Bolyai Univ, Fac Math & Comp Sci, 1 Mihail Kogalniceanu Str, RO-400084 Cluj Napoca, Romania
关键词
Multi -agent systems; Reinforcement learning; Deep Q -learning;
D O I
10.1016/j.neucom.2023.126507
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We model cooperative-competitive social group dynamics with multi-agent environments, specialized in cases with a large number of agents from only a few distinct types. The multi-agent optimization problems are addressed in turn with multi-agent reinforcement learning algorithms to obtain flexible and robust solutions. We analyze the effectiveness of centralized and decentralized algorithms using three variants of deep Q-networks on these cooperative-competitive environments: first, we use the decentralized training independent learning with deep Q-networks, secondly the centralized monotonic value factorizations for deep learning, and lastly the multi-agent variational exploration. We test the algorithms in simulated predator-prey multi-agent environments in two distinct environments: the adversary pursuit and simple tag. The experiments highlight the performance of the different deep Q-learning methods, and we conclude that decentralized training of deep Q-networks accumulates higher episode rewards during training and evaluation in comparison with the selected centralized learning approaches.& COPY; 2023 Elsevier B.V. All rights reserved.
引用
收藏
页数:8
相关论文
共 50 条
  • [41] NeuroHex: A Deep Q-learning Hex Agent
    Young, Kenny
    Vasan, Gautham
    Hayward, Ryan
    COMPUTER GAMES: 5TH WORKSHOP ON COMPUTER GAMES, CGW 2016, AND 5TH WORKSHOP ON GENERAL INTELLIGENCE IN GAME-PLAYING AGENTS, GIGA 2016, HELD IN CONJUNCTION WITH THE 25TH INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2016, NEW YORK, USA, JULY 9-10, 2016, 2017, 705 : 3 - 18
  • [42] Evolution of cooperation in a mixed cooperative-competitive structured population
    Lyu, Ding
    Liu, Hanxiao
    Wang, Lin
    Wang, Xiaofan
    PHYSICA A-STATISTICAL MECHANICS AND ITS APPLICATIONS, 2024, 652
  • [43] Evaluating semi-cooperative Nash/Stackelberg Q-learning for traffic routes plan in a single intersection
    Guo, Jian
    Harmati, Istvan
    CONTROL ENGINEERING PRACTICE, 2020, 102
  • [44] Learning Automata Based Q-Learning for Content Placement in Cooperative Caching
    Yang, Zhong
    Liu, Yuanwei
    Chen, Yue
    Jiao, Lei
    IEEE TRANSACTIONS ON COMMUNICATIONS, 2020, 68 (06) : 3667 - 3680
  • [45] Evolutionary Dynamics of Q-Learning over the Sequence
    Panozzo, Fabio
    Gatti, Nicola
    Restelli, Marcello
    PROCEEDINGS OF THE TWENTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2014, : 2034 - 2040
  • [46] Demand response model: A cooperative-competitive multi-agent reinforcement learning approach
    Salazar, Eduardo J.
    Rosero, Veronica
    Gabrielski, Jawana
    Samper, Mauricio E.
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2024, 133
  • [47] QLP: Deep Q-Learning for Pruning Deep Neural Networks
    Camci, Efe
    Gupta, Manas
    Wu, Min
    Lin, Jie
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (10) : 6488 - 6501
  • [48] GENDER DIFFERENCE OF SOCIAL BEHAVIOR IN THE COOPERATIVE-COMPETITIVE GAME
    Hong, Jon-Chao
    Hwang, Ming-Yueh
    Peng, Yu-Chi
    12TH INTERNATIONAL EDUCATIONAL TECHNOLOGY CONFERENCE - IETC 2012, 2012, 64 : 244 - 254
  • [49] Controlling the cooperative-competitive boundary in niched genetic algorithms
    Horn, J
    GECCO-99: PROCEEDINGS OF THE GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE, 1999, : 305 - 312
  • [50] Evaluating Deep Q-Learning Algorithms for Controlling Blood Glucose in In Silico Type 1 Diabetes
    Tejedor, Miguel
    Hjerde, Sigurd Nordtveit
    Myhre, Jonas Nordhaug
    Godtliebsen, Fred
    DIAGNOSTICS, 2023, 13 (19)