Control of a Water Tank System with Value Function Approximation

被引:0
作者
Lalvani, Shamal [1 ]
Katsaggelos, Aggelos [1 ]
机构
[1] Northwestern Univ, Evanston, IL 60208 USA
来源
ARTIFICIAL INTELLIGENCE APPLICATIONS AND INNOVATIONS, AIAI 2023, PT I | 2023年 / 675卷
关键词
Reinforcement Learning; Optimal Control; Water Tank System;
D O I
10.1007/978-3-031-34111-3_4
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We consider a system of two identical rectangular shaped water tanks. A source of constant water inflow is available, which may only be directed to one tank at a time. The objective is to find a control policy to maximize the final sum of the water levels at some terminal time T, subject to minimum water level constraints on each tank. Water exits each tank corresponding to Toricelli's law (i.e., the velocity depends on the current water level). We derive a closed form dynamic programming solution in discrete time to this problem without the water-level threshold constraints. Subsequently, we implement the value iteration algorithm on a set of support points to find a control policy with the threshold constraints, where a random forest regressor is iteratively used to update the value function. Our results show consistency between the dynamic programming solution and the value iteration solution.
引用
收藏
页码:36 / 44
页数:9
相关论文
共 50 条
[31]   A New Approach for Value Function Approximation Based on Automatic State Partition [J].
Zeng, Jiaan ;
Han, Yinghua .
IMECS 2009: INTERNATIONAL MULTI-CONFERENCE OF ENGINEERS AND COMPUTER SCIENTISTS, VOLS I AND II, 2009, :208-213
[32]   Hybrid Modeling and Optimal Control of a Two-Tank System as a Switched System [J].
Mahboubi, H. ;
Moshiri, B. ;
Seddigh, A. Khaki .
PROCEEDINGS OF WORLD ACADEMY OF SCIENCE, ENGINEERING AND TECHNOLOGY, VOL 11, 2006, 11 :139-+
[33]   Discrete-time optimal control of electric hot water tank [J].
Beeker, Nathanael ;
Malisani, Paul ;
Petit, Nicolas .
IFAC PAPERSONLINE, 2016, 49 (07) :882-888
[34]   The value function of singularly perturbed control systems [J].
Artstein, Z ;
Gaitsgory, V .
APPLIED MATHEMATICS AND OPTIMIZATION, 2000, 41 (03) :425-445
[35]   Local regularity of the value function in optimal control [J].
Cannarsa, P. ;
Frankowska, H. .
SYSTEMS & CONTROL LETTERS, 2013, 62 (09) :791-794
[36]   Data-driven optimal control of wind turbines using reinforcement learning with function approximation [J].
Peng, Shenglin ;
Feng, Qianmei .
COMPUTERS & INDUSTRIAL ENGINEERING, 2023, 176
[37]   Low-rank State-action Value-function Approximation [J].
Rozada, Sergio ;
Tenorio, Victor ;
Marques, Antonio G. .
29TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO 2021), 2021, :1471-1475
[38]   Efficient exploration through active learning for value function approximation in reinforcement learning [J].
Akiyama, Takayuki ;
Hachiya, Hirotaka ;
Sugiyama, Masashi .
NEURAL NETWORKS, 2010, 23 (05) :639-648
[39]   Nonlinear Value Function Approximation Method With Easy Hyperparameter Tuning and Convergence Guarantee [J].
Watabe, Yuuya ;
Shibuya, Takeshi .
IEEE ACCESS, 2025, 13 :30117-30126
[40]   Least Absolute Policy Iteration-A Robust Approach to Value Function Approximation [J].
Sugiyama, Masashi ;
Hachiya, Hirotaka ;
Kashima, Hisashi ;
Morimura, Tetsuro .
IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2010, E93D (09) :2555-2565