Effect of Q-learning on the evolution of cooperation behavior in collective motion: An improved Vicsek model

被引:3
作者
Wang, Chengjie [1 ,2 ]
Deng, Juan [3 ]
Zhao, Hui [4 ]
Li, Li [1 ,2 ]
机构
[1] Tongji Univ, Dept Control Sci & Engn, Coll Elect & Informat Engn, 4800 Caoan Highway, Shanghai 201804, Peoples R China
[2] Shanghai Res Inst Intelligent Autonomous Syst, Shanghai 200120, Peoples R China
[3] Natl Univ Def Technol, Coll Sci, Changsha 410073, Peoples R China
[4] Tongji Univ, Coll Elect & Informat Engn, Dept Comp & Sci, 4800 Caoan Highway, Shanghai 201804, Peoples R China
基金
中国国家自然科学基金;
关键词
Vicsek model; Evolutionary game theory; Q-learning; Cooperation; DYNAMICS;
D O I
10.1016/j.amc.2024.128956
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
There have been numerous studies on collective behavior, among which communication between agents can have a great impact on both the payoff and the cost of making decisions. Research usually focuses on how to improve the collective synchronization rate or accelerate the process of cooperation under given communication cost constraints. In this context, evolutionary game theory (EGT) and reinforcement learning (RL) arise as essential frameworks for tackling this intricate problem. In this study, an adapted Vicsek model is introduced, wherein agents exhibit varying movement patterns contingent on their chosen strategies. Each agent gains a payoff determined by the advantages of collective motion juxtaposed with the cost of communicating with neighboring agents. Individuals choose the objective agents based on the Q-learning strategy and then adapt their strategies following the Fermi rule. The research reveals that the utmost level of cooperation and synchronization can be attained at an optimal communication radius after applying Q-learning. Similar conclusions have been drawn from research on the influence of random noise and relative cost. Different cost functions were considered in the study to demonstrate the robustness of the proposed model and conclusions under a wide range of conditions. ( https://github .com /WangchengjieT /VM-EGT-Q) )
引用
收藏
页数:11
相关论文
共 35 条
[1]   Coevolution of Synchronization and Cooperation in Costly Networked Interactions [J].
Antonioni, Alberto ;
Cardillo, Alessio .
PHYSICAL REVIEW LETTERS, 2017, 118 (23)
[2]   Evolution of cooperation in social dilemmas under the coexistence of aspiration and imitation mechanisms [J].
Arefin, Md Rajib ;
Tanimoto, Jun .
PHYSICAL REVIEW E, 2020, 102 (03)
[3]   Evolutionary Dynamics of Multi-Agent Learning: A Survey [J].
Bloembergen, Daan ;
Tuyls, Karl ;
Hennes, Daniel ;
Kaisers, Michael .
JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH, 2015, 53 :659-697
[4]   CAUCHY THEORY FOR GENERAL KINETIC VICSEK MODELS IN COLLECTIVE DYNAMICS AND MEAN-FIELD LIMIT APPROXIMATIONS [J].
Briant, Marc ;
Diez, Antoine ;
Merino-Aceituno, Sara .
SIAM JOURNAL ON MATHEMATICAL ANALYSIS, 2022, 54 (01) :1131-1168
[5]   Flocking of two unfriendly species: The two-species Vicsek model [J].
Chatterjee, Swarnajit ;
Mangeat, Matthieu ;
Woo, Chul-Ung ;
Rieger, Heiko ;
Noh, Jae Dong .
PHYSICAL REVIEW E, 2023, 107 (02)
[6]  
Darwin C, 2009, ON THE ORIGIN OF SPECIES, P1, DOI 10.1017/CBO9780511694295.004
[7]   Q-learning boosts the evolution of cooperation in structured population by involving extortion [J].
Ding, Hong ;
Zhang, Geng-shun ;
Wang, Shi-hao ;
Li, Juan ;
Wang, Zhen .
PHYSICA A-STATISTICAL MECHANICS AND ITS APPLICATIONS, 2019, 536
[8]   Reinforcement learning explains various conditional cooperation [J].
Geng, Yini ;
Liu, Yifan ;
Lu, Yikang ;
Shen, Chen ;
Shi, Lei .
APPLIED MATHEMATICS AND COMPUTATION, 2022, 427
[9]   Mercenary punishment in structured populations [J].
Lee, Hsuan-Wei ;
Cleveland, Colin ;
Szolnoki, Attila .
APPLIED MATHEMATICS AND COMPUTATION, 2022, 417
[10]   Changing the Intensity of Interaction Based on Individual Behavior in the Iterated Prisoner's Dilemma Game [J].
Li, Jiaqi ;
Zhang, Chunyan ;
Sun, Qinglin ;
Chen, Zengqiang ;
Zhang, Jianlei .
IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, 2017, 21 (04) :506-517