Fish growth trajectory tracking using Q-learning in precision aquaculture

被引:16
作者
Chahid, Abderrazak [1 ]
N'Doye, Ibrahima [1 ]
Majoris, John E. [2 ]
Berumen, Michael L. [2 ]
Laleg-Kirati, Taous-Meriem [1 ]
机构
[1] King Abdullah Univ Sci & Technol KAUST, Elect & Math Sci & Engn Div CEMSE, Thuwal 239556900, Makkah Province, Saudi Arabia
[2] King Abdullah Univ Sci & Technol KAUST, Red Sea Res Ctr, Biol & Environm Sci & Engn Div, Thuwal 239556900, Makkah Province, Saudi Arabia
关键词
Fish growth model; Reference trajectory tracking; Markov decision process; Process control; Q-learning; Reinforcement learning; TILAPIA OREOCHROMIS-NILOTICUS; MODEL; OUTPUT; POND;
D O I
10.1016/j.aquaculture.2021.737838
中图分类号
S9 [水产、渔业];
学科分类号
0908 ;
摘要
This paper studies fish growth trajectory tracking using Q-learning under a representative bioenergetic growth model of Nile tilapia (Oreochromis niloticus). The fish growth rate varies in practice and cannot be easily estimated due to the complex aquaculture condition and variable environmental factors. Additionally, the growth trajectory tracking problem is challenging to solve by most of the model-based control approaches due to the nonlinear couplings and interactions between multi-inputs such as temperature, dissolved oxygen, un-ionized ammonia, and the model uncertainty of the fish growth system. We formulate the growth trajectory tracking problem as sampled-data optimal control using discrete state-action pairs Markov decision process on the simulated growth trajectories data to mimic the real aquaculture environment adequately. We propose two Q-learning algorithms that learn the optimal control policy from the simulated data of the fish growth trajectories beginning from the juvenile stage until the desired market weight in the aquaculture environment. The first Q-learning scheme learns the optimal feeding control policy to fish growth rate cultured in cages, while the second one online updates the optimal feeding control policy within an optimal temperature profile for the aquaculture fish growth rate in tanks. The simulation results demonstrate that both Q-learning control strategies achieve good trajectory tracking performance with lower feeding rates and help compensate for the environmental changes of the manipulated variables and the bioenergetic model uncertainties of fish growth in the aquaculture environment. The proposed Q-learning control policies achieve 1.7% and 6.6% relative trajectory tracking errors of the average total weight of fish from both tanks on land and floating cages, respectively. Furthermore, the feeding and temperature control policies reduce 11% relative feeding quantity of the food waste in tanks on land compared to the floating cages where the water temperature is maintained at the ambient temperature of 29.7 degrees C.
引用
收藏
页数:9
相关论文
共 50 条
  • [41] Using the ITS Components in Improving the Q-Learning Policy for Instructional Sequencing
    Yessad, Amel
    AUGMENTED INTELLIGENCE AND INTELLIGENT TUTORING SYSTEMS, ITS 2023, 2023, 13891 : 247 - 256
  • [42] Robust Attitude Control of an Agile Aircraft Using Improved Q-Learning
    Zahmatkesh, Mohsen
    Emami, Seyyed Ali
    Banazadeh, Afshin
    Castaldi, Paolo
    ACTUATORS, 2022, 11 (12)
  • [43] The acquisition of sociality by using Q-learning in a multi-agent environment
    Nagayuki, Yasuo
    PROCEEDINGS OF THE SIXTEENTH INTERNATIONAL SYMPOSIUM ON ARTIFICIAL LIFE AND ROBOTICS (AROB 16TH '11), 2011, : 820 - 823
  • [44] Simulating SQL injection vulnerability exploitation using Q-learning reinforcement learning agents
    Erdodi, Laszlo
    Sommervoll, Avald Aslaugson
    Zennaro, Fabio Massimo
    JOURNAL OF INFORMATION SECURITY AND APPLICATIONS, 2021, 61
  • [45] Incentive Reward for Efficient WiFi Offloading using Q-Learning Approach
    Fakhfakh, Emna
    Hamouda, Soumaya
    2017 13TH INTERNATIONAL WIRELESS COMMUNICATIONS AND MOBILE COMPUTING CONFERENCE (IWCMC), 2017, : 1114 - 1119
  • [46] Solving a Job Shop Scheduling Problem Using Q-Learning Algorithm
    Belmamoune, Manal Abir
    Ghomri, Latefa
    Yahouni, Zakaria
    12TH INTERNATIONAL WORKSHOP ON SERVICE ORIENTED, HOLONIC AND MULTI-AGENT MANUFACTURING SYSTEMS FOR INDUSTRY OF THE FUTURE, SOHOMA 2022, 2023, 1083 : 196 - 209
  • [47] Efficient Q-learning hyperparameter tuning using FOX optimization algorithm
    Jumaah, Mahmood A.
    Ali, Yossra H.
    Rashid, Tarik A.
    RESULTS IN ENGINEERING, 2025, 25
  • [48] An intelligent financial portfolio trading strategy using deep Q-learning
    Park, Hyungjun
    Sim, Min Kyu
    Choi, Dong Gu
    EXPERT SYSTEMS WITH APPLICATIONS, 2020, 158 (158)
  • [49] A Hand Gesture Recognition System Using EMG and Reinforcement Learning: A Q-Learning Approach
    Vasconez, Juan Pablo
    Barona Lopez, Lorena Isabel
    Valdivieso Caraguay, Angel Leonardo
    Cruz, Patricio J.
    Alvarez, Robin
    Benalcazar, Marco E.
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2021, PT IV, 2021, 12894 : 580 - 591
  • [50] Accelerated multi-objective task learning using modified Q-learning algorithm
    Rajamohan, Varun Prakash
    Jagatheesaperumal, Senthil Kumar
    INTERNATIONAL JOURNAL OF AD HOC AND UBIQUITOUS COMPUTING, 2024, 47 (01) : 28 - 37