Control of a Water Tank System with Value Function Approximation

被引:0
作者
Lalvani, Shamal [1 ]
Katsaggelos, Aggelos [1 ]
机构
[1] Northwestern Univ, Evanston, IL 60208 USA
来源
ARTIFICIAL INTELLIGENCE APPLICATIONS AND INNOVATIONS, AIAI 2023, PT I | 2023年 / 675卷
关键词
Reinforcement Learning; Optimal Control; Water Tank System;
D O I
10.1007/978-3-031-34111-3_4
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We consider a system of two identical rectangular shaped water tanks. A source of constant water inflow is available, which may only be directed to one tank at a time. The objective is to find a control policy to maximize the final sum of the water levels at some terminal time T, subject to minimum water level constraints on each tank. Water exits each tank corresponding to Toricelli's law (i.e., the velocity depends on the current water level). We derive a closed form dynamic programming solution in discrete time to this problem without the water-level threshold constraints. Subsequently, we implement the value iteration algorithm on a set of support points to find a control policy with the threshold constraints, where a random forest regressor is iteratively used to update the value function. Our results show consistency between the dynamic programming solution and the value iteration solution.
引用
收藏
页码:36 / 44
页数:9
相关论文
共 50 条
[41]   Optimization control of the double-capacity water tank-level system using the deep deterministic policy gradient algorithm [J].
Ye, Likun ;
Jiang, Pei .
ENGINEERING REPORTS, 2023, 5 (11)
[42]   Continuity of the Value Function for Stochastic Sparse Optimal Control [J].
Ito, Kaito ;
Ikeda, Takuya ;
Kashima, Kenji .
IFAC PAPERSONLINE, 2020, 53 (02) :7179-7184
[43]   On singularities of value function for Bolza optimal control problem [J].
Frankowska, H ;
Ochal, A .
JOURNAL OF MATHEMATICAL ANALYSIS AND APPLICATIONS, 2005, 306 (02) :714-729
[44]   Reinforcement learning control with function approximation via multivariate simplex splines [J].
Feng, Yiting ;
Zhou, Ye ;
Ho, Hann Woei ;
Mat Isa, Nor Ashidi .
INTERNATIONAL JOURNAL OF ADAPTIVE CONTROL AND SIGNAL PROCESSING, 2023,
[45]   A review of stochastic algorithms with continuous value function approximation and some new approximate policy iteration algorithms for multidimensional continuous applications [J].
Warren B.POWELL .
JournalofControlTheoryandApplications, 2011, 9 (03) :336-352
[46]   A review of stochastic algorithms with continuous value function approximation and some new approximate policy iteration algorithms for multidimensional continuous applications [J].
Powell W.B. ;
Ma J. .
Journal of Control Theory and Applications, 2011, 9 (3) :336-352
[47]   Dynamic Spectrum Anti-Jamming With Reinforcement Learning Based on Value Function Approximation [J].
Zhu, Xinyu ;
Huang, Yang ;
Wang, Shaoyu ;
Wu, Qihui ;
Ge, Xiaohu ;
Liu, Yuan ;
Gao, Zhen .
IEEE WIRELESS COMMUNICATIONS LETTERS, 2023, 12 (02) :386-390
[48]   A Novel Artificial Hydrocarbon Networks Based Value Function Approximation in Hierarchical Reinforcement Learning [J].
Ponce, Hiram .
ADVANCES IN SOFT COMPUTING, MICAI 2016, PT II, 2017, 10062 :211-225
[49]   Tensor and Matrix Low-Rank Value-Function Approximation in Reinforcement Learning [J].
Rozada, Sergio ;
Paternain, Santiago ;
Marques, Antonio .
IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2024, 72 :1634-1649
[50]   GEODESIC DISTANCE-BASED KERNEL CONSTRUCTION FOR GAUSSIAN PROCESS VALUE FUNCTION APPROXIMATION [J].
Jakab, Hunor .
KEPT 2011: KNOWLEDGE ENGINEERING PRINCIPLES AND TECHNIQUES, 2011, :315-324