Evaluating cooperative-competitive dynamics with deep Q-learning

被引:2
|
作者
Kopacz, Aniko [1 ]
Csato, Lehel [1 ]
Chira, Camelia [1 ]
机构
[1] Babes Bolyai Univ, Fac Math & Comp Sci, 1 Mihail Kogalniceanu Str, RO-400084 Cluj Napoca, Romania
关键词
Multi -agent systems; Reinforcement learning; Deep Q -learning;
D O I
10.1016/j.neucom.2023.126507
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We model cooperative-competitive social group dynamics with multi-agent environments, specialized in cases with a large number of agents from only a few distinct types. The multi-agent optimization problems are addressed in turn with multi-agent reinforcement learning algorithms to obtain flexible and robust solutions. We analyze the effectiveness of centralized and decentralized algorithms using three variants of deep Q-networks on these cooperative-competitive environments: first, we use the decentralized training independent learning with deep Q-networks, secondly the centralized monotonic value factorizations for deep learning, and lastly the multi-agent variational exploration. We test the algorithms in simulated predator-prey multi-agent environments in two distinct environments: the adversary pursuit and simple tag. The experiments highlight the performance of the different deep Q-learning methods, and we conclude that decentralized training of deep Q-networks accumulates higher episode rewards during training and evaluation in comparison with the selected centralized learning approaches.& COPY; 2023 Elsevier B.V. All rights reserved.
引用
收藏
页数:8
相关论文
共 50 条
  • [11] Deep Reinforcement Learning with Double Q-Learning
    van Hasselt, Hado
    Guez, Arthur
    Silver, David
    THIRTIETH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2016, : 2094 - 2100
  • [12] Cooperative Q-learning: the knowledge sharing issue
    Ahmadabadi, MN
    Asadpour, M
    Nakano, E
    ADVANCED ROBOTICS, 2001, 15 (08) : 815 - 832
  • [13] Cooperative Q-Learning Based on Maturity of the Policy
    Yang, Mao
    Tian, Yantao
    Liu, Xiaomei
    2009 IEEE INTERNATIONAL CONFERENCE ON MECHATRONICS AND AUTOMATION, VOLS 1-7, CONFERENCE PROCEEDINGS, 2009, : 1352 - 1356
  • [14] Comparison of Deep Q-Learning, Q-Learning and SARSA Reinforced Learning for Robot Local Navigation
    Anas, Hafiq
    Ong, Wee Hong
    Malik, Owais Ahmed
    ROBOT INTELLIGENCE TECHNOLOGY AND APPLICATIONS 6, 2022, 429 : 443 - 454
  • [15] Hierarchical clustering with deep Q-learning
    Forster, Richard
    Fulop, Agnes
    ACTA UNIVERSITATIS SAPIENTIAE INFORMATICA, 2018, 10 (01) : 86 - 109
  • [16] Active deep Q-learning with demonstration
    Si-An Chen
    Voot Tangkaratt
    Hsuan-Tien Lin
    Masashi Sugiyama
    Machine Learning, 2020, 109 : 1699 - 1725
  • [17] Active deep Q-learning with demonstration
    Chen, Si-An
    Tangkaratt, Voot
    Lin, Hsuan-Tien
    Sugiyama, Masashi
    MACHINE LEARNING, 2020, 109 (9-10) : 1699 - 1725
  • [18] A Theoretical Analysis of Deep Q-Learning
    Fan, Jianqing
    Wang, Zhaoran
    Xie, Yuchen
    Yang, Zhuoran
    LEARNING FOR DYNAMICS AND CONTROL, VOL 120, 2020, 120 : 486 - 489
  • [19] Deep Q-Learning from Demonstrations
    Hester, Todd
    Vecerik, Matej
    Pietquin, Olivier
    Lanctot, Marc
    Schaul, Tom
    Piot, Bilal
    Horgan, Dan
    Quan, John
    Sendonaris, Andrew
    Osband, Ian
    Dulac-Arnold, Gabriel
    Agapiou, John
    Leibo, Joel Z.
    Gruslys, Audrunas
    THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 3223 - 3230
  • [20] An Online Home Energy Management System using Q-Learning and Deep Q-Learning
    Izmitligil, Hasan
    Karamancioglu, Abdurrahman
    SUSTAINABLE COMPUTING-INFORMATICS & SYSTEMS, 2024, 43