Learning obstacle avoidance and predation in complex reef environments with deep reinforcement learning

被引:4
作者
Hou, Ji [1 ,2 ]
He, Changling [1 ,2 ]
Li, Tao [1 ,2 ]
Zhang, Chunze [1 ,2 ]
Zhou, Qin [3 ]
机构
[1] Chongqing Jiaotong Univ, Southwest Res Inst Hydraul & Water Transport Engn, Chongqing 402247, Peoples R China
[2] Chongqing Jiaotong Univ, Coll River & Ocean Engn, Chongqing 400074, Peoples R China
[3] Chongqing Xike Consulting Co LTD Water Transport E, Chongqing 402247, Peoples R China
基金
中国国家自然科学基金;
关键词
deep reinforcement learning; fluid-structure interaction; immersed boundary lattice Boltzmann method; intelligent fish; sparse reward; BOLTZMANN COUPLING SCHEME; IMPROVEMENT; STABILITY; FISH;
D O I
10.1088/1748-3190/ad6544
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
The reef ecosystem plays a vital role as a habitat for fish species with limited swimming capabilities, serving not only as a sanctuary and food source but also influencing their behavioral tendencies. Understanding the intricate mechanism through which fish adeptly navigate the moving targets within reef environments within complex water flow, all while evading obstacles and maintaining stable postures, has remained a challenging and prominent subject in the realms of fish behavior, ecology, and biomimetics alike. An integrated simulation framework is used to investigate fish predation problems within intricate environments, combining deep reinforcement learning algorithms (DRL) with high-precision fluid-structure interaction numerical methods-immersed boundary lattice Boltzmann method (lB-LBM). The Soft Actor-Critic (SAC) algorithm is used to improve the intelligent fish's capacity for random exploration, tackling the multi-objective sparse reward challenge inherent in real-world scenarios. Additionally, a reward shaping method tailored to its action purposes has been developed, capable of capturing outcomes and trend characteristics effectively. The convergence and robustness advantages of the method elucidated in this paper are showcased through two case studies: one addressing fish capturing randomly moving targets in hydrostatic flow field, and the other focusing on fish counter-current foraging in reef environments to capture drifting food. A comprehensive analysis was conducted of the influence and significance of various reward types on the decision-making processes of intelligent fish within intricate environments.
引用
收藏
页数:15
相关论文
共 26 条
[1]   Optimal Control of Point-to-Point Navigation in Turbulent Time Dependent Flows Using Reinforcement Learning [J].
Buzzicotti, Michele ;
Biferale, Luca ;
Bonaccorso, Fabio ;
Di Leoni, Patricio Clark ;
Gustavsson, Kristian .
AIXIA 2020 - ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 12414 :223-234
[2]   Flow Navigation by Smart Microswimmers via Reinforcement Learning [J].
Colabrese, Simona ;
Gustavsson, Kristian ;
Celani, Antonio ;
Biferale, Luca .
PHYSICAL REVIEW LETTERS, 2017, 118 (15)
[3]   Simulation of hydraulic characteristics of an inclined overflow gate by the free-surface lattice Boltzmann-immersed boundary coupling scheme [J].
Diao, Wei ;
Cheng, Yongguang ;
Xu, Manjun ;
Wu, Jiayang ;
Zhang, Chunze ;
Zhou, Wei .
ENGINEERING APPLICATIONS OF COMPUTATIONAL FLUID MECHANICS, 2017, 12 (01) :250-260
[4]   A review on deep reinforcement learning for fluid mechanics [J].
Garnier, Paul ;
Viquerat, Jonathan ;
Rabault, Jean ;
Larcher, Aurelien ;
Kuhnle, Alexander ;
Hachem, Elie .
COMPUTERS & FLUIDS, 2021, 225 (225)
[5]  
Haarnoja T, 2018, PR MACH LEARN RES, V80
[7]   Fish can save energy via proprioceptive sensing [J].
Li, Liang ;
Liu, Danshi ;
Deng, Jian ;
Lutz, Matthew J. ;
Xie, Guangming .
BIOINSPIRATION & BIOMIMETICS, 2021, 16 (05)
[8]   Biological Inspiration: From Carangiform Fish to Multi-Joint Robotic Fish [J].
Liu, Jindong ;
Hu, Huosheng .
JOURNAL OF BIONIC ENGINEERING, 2010, 7 (01) :35-48
[9]   A Survey of Reinforcement Learning Algorithms for Dynamically Varying Environments [J].
Padakandla, Sindhu .
ACM COMPUTING SURVEYS, 2021, 54 (06)
[10]   Deep reinforcement learning in fluid mechanics: A promising method for both active flow control and shape optimization [J].
Rabault, Jean ;
Ren, Feng ;
Zhang, Wei ;
Tang, Hui ;
Xu, Hui .
JOURNAL OF HYDRODYNAMICS, 2020, 32 (02) :234-246