Emergence of chemotactic strategies with multi-agent reinforcement learning

被引:0
|
作者
Tovey, Samuel [1 ]
Lohrmann, Christoph [1 ]
Holm, Christian [1 ]
机构
[1] Univ Stuttgart, Inst Computat Phys, D-70569 Stuttgart, Germany
来源
MACHINE LEARNING-SCIENCE AND TECHNOLOGY | 2024年 / 5卷 / 03期
关键词
reinforcement learning; microrobotics; chemotaxis; active matter; biophysics; MOTION;
D O I
10.1088/2632-2153/ad5f73
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Reinforcement learning (RL) is a flexible and efficient method for programming micro-robots in complex environments. Here we investigate whether RL can provide insights into biological systems when trained to perform chemotaxis. Namely, whether we can learn about how intelligent agents process given information in order to swim towards a target. We run simulations covering a range of agent shapes, sizes, and swim speeds to determine if the physical constraints on biological swimmers, namely Brownian motion, lead to regions where reinforcement learners' training fails. We find that the RL agents can perform chemotaxis as soon as it is physically possible and, in some cases, even before the active swimming overpowers the stochastic environment. We study the efficiency of the emergent policy and identify convergence in agent size and swim speeds. Finally, we study the strategy adopted by the RL algorithm to explain how the agents perform their tasks. To this end, we identify three emerging dominant strategies and several rare approaches taken. These strategies, whilst producing almost identical trajectories in simulation, are distinct and give insight into the possible mechanisms behind which biological agents explore their environment and respond to changing conditions.
引用
收藏
页数:20
相关论文
共 50 条
  • [31] Exploration Strategies for Learning in Multi-agent Foraging
    Mohan, Yogeswaran
    Ponnambalam, S. G.
    SWARM, EVOLUTIONARY, AND MEMETIC COMPUTING, PT II, 2011, 7077 : 17 - 26
  • [32] Multi-Agent Reinforcement Learning With Decentralized Distribution Correction
    Li, Kuo
    Jia, Qing-Shan
    IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2025, 22 : 1684 - 1696
  • [33] Multi-Agent Deep Reinforcement Learning with Emergent Communication
    Simoes, David
    Lau, Nuno
    Reis, Luis Paulo
    2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,
  • [34] Experience Selection in Multi-Agent Deep Reinforcement Learning
    Wang, Yishen
    Zhang, Zongzhang
    2019 IEEE 31ST INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI 2019), 2019, : 864 - 870
  • [35] Shapley Counterfactual Credits for Multi-Agent Reinforcement Learning
    Li, Jiahui
    Kuang, Kun
    Wang, Baoxiang
    Liu, Furui
    Chen, Long
    Wu, Fei
    Xiao, Jun
    KDD '21: PROCEEDINGS OF THE 27TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2021, : 934 - 942
  • [36] Towards reinforcement learning for holonic multi-agent systems
    Abdoos, Monireh
    Mozayani, Nasser
    Bazzan, Ana L. C.
    INTELLIGENT DATA ANALYSIS, 2015, 19 (02) : 211 - 232
  • [37] Battlefield Environment Design for Multi-agent Reinforcement Learning
    Do, Seungwon
    Baek, Jaeuk
    Jun, Sungwoo
    Lee, Changeun
    2022 IEEE INTERNATIONAL CONFERENCE ON BIG DATA AND SMART COMPUTING (IEEE BIGCOMP 2022), 2022, : 318 - 319
  • [38] Shaping multi-agent systems with gradient reinforcement learning
    Olivier Buffet
    Alain Dutech
    François Charpillet
    Autonomous Agents and Multi-Agent Systems, 2007, 15 : 197 - 220
  • [39] Generating Multi-agent Patrol Areas by Reinforcement Learning
    Park, Bumjin
    Kang, Cheongwoong
    Choi, Jaesik
    2021 21ST INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND SYSTEMS (ICCAS 2021), 2021, : 104 - 107
  • [40] Rationality of reward sharing in multi-agent reinforcement learning
    Miyazaki, K
    Kobayashi, S
    NEW GENERATION COMPUTING, 2001, 19 (02) : 157 - 172