Reinforcement learning control method for real-time hybrid simulation based on deep deterministic policy gradient algorithm

被引:13
作者
Li, Ning [1 ,3 ,4 ]
Tang, Jichuan [1 ,2 ]
Li, Zhong-Xian [1 ,3 ,4 ]
Gao, Xiuyu [5 ]
机构
[1] Tianjin Univ, Sch Civil Engn, Tianjin 300350, Peoples R China
[2] Nanyang Technol Univ, Sch Civil & Environm Engn, Singapore, Singapore
[3] Tianjin Univ, Minist Educ, Key Lab Coast Civil Struct Safety, Tianjin, Peoples R China
[4] Tianjin Univ, Key Lab Earthquake Engn Simulat & Seism Resilienc, China Earthquake Adm, Tianjin, Peoples R China
[5] MTS Corp, Eden Prairie, MN USA
基金
国家重点研发计划;
关键词
deep deterministic policy gradient algorithm; hybrid control; real-time hybrid simulation; reinforcement learning; underwater shaking table; ACTUATOR CONTROL; COMPENSATION; PERFORMANCE;
D O I
10.1002/stc.3035
中图分类号
TU [建筑科学];
学科分类号
0813 ;
摘要
The tracking performance of an actuation transfer system in a real-time hybrid simulation (RTHS) frequently faces accuracy and robustness challenges under constraints and complicated environments with uncertainties. This study proposes a novel control approach based on the deep deterministic policy gradient algorithm in reinforcement learning (RL) combined with feedforward (FF) compensation, which emphasizes the implementation of shaking table control and substructure RTHS. The proposed method first describes the control plant within the RL environment. Then, the agent is trained offline to develop optimized control policies for interaction with the environment. A series of validation tests were conducted to assess the performance of the proposed method, starting with the dynamic testing of underwater shaking table control and then a virtual RTHS benchmark problem. For complex systems, such as controlling the underwater shaking table, the proposed algorithm, FF, and adaptive time series (ATS) compensation methods are compared under various water depths and motions. The results show better performance and wider broadband frequency applicability under different shaking table dynamic-coupling effects. Next, a controller based on the proposed method was designed by extending the virtual RTHS via the configuration of the control plant and substructure division, as provided in the RTHS benchmark problem. The proposed RL controller also improved the tracking accuracy and robustness of conventional FF compensators against unmodeled dynamics and perturbation uncertainties. This controller can be extended to further advanced control strategies as a component of model-based control methods.
引用
收藏
页数:24
相关论文
共 43 条
[1]   A Deep Reinforcement Learning Approach to Improve the Learning Performance in Process Control [J].
Bao, Yaoyao ;
Zhu, Yuanming ;
Qian, Feng .
INDUSTRIAL & ENGINEERING CHEMISTRY RESEARCH, 2021, 60 (15) :5504-5515
[2]  
Berthelsen PA, 2016, PROCEEDINGS OF THE ASME 35TH INTERNATIONAL CONFERENCE ON OCEAN, OFFSHORE AND ARCTIC ENGINEERING , 2016, VOL 6
[3]  
Bertsekas D.P., 2019, Reinforcement Learning and Optimal Control
[4]  
Brunton S L., 2019, Data-Driven Science and Engineering, DOI 10.1017/9781108380690
[5]  
Carrion J.E., 2007, MODEL BASED STRATEGI
[6]   Adaptive time series compensator for delay compensation of servo-hydraulic actuator systems for real-time hybrid simulation [J].
Chae, Yunbyeong ;
Kazemibidokhti, Karim ;
Ricles, James M. .
EARTHQUAKE ENGINEERING & STRUCTURAL DYNAMICS, 2013, 42 (11) :1697-1715
[7]  
Chen C., 2007, Development and numerical simulation of hybrid effective force testing method
[8]   Adaptive model-based tracking control for real-time hybrid simulation [J].
Chen, Pei-Ching ;
Chang, Chia-Ming ;
Spencer, Billie F., Jr. ;
Tsai, Keh-Chyuan .
BULLETIN OF EARTHQUAKE ENGINEERING, 2015, 13 (06) :1633-1653
[9]   Stability and delay compensation for real-time substructure testing [J].
Darby, AP ;
Williams, MS ;
Blakeborough, A .
JOURNAL OF ENGINEERING MECHANICS, 2002, 128 (12) :1276-1284
[10]   Application of model-based compensation methods to real-time hybrid simulation benchmark [J].
Fermandois, Gaston A. .
MECHANICAL SYSTEMS AND SIGNAL PROCESSING, 2019, 131 :394-416