Proximal policy optimization-based controller for chaotic systems

被引:6
作者
Yau, Her-Terng [1 ,2 ,3 ]
Kuo, Ping-Huan [1 ,2 ]
Luan, Po-Chien [2 ]
Tseng, Yung-Ruen [2 ]
机构
[1] Natl Chung Cheng Univ, Dept Mech Engn, Chiayi, Taiwan
[2] Natl Chung Cheng Univ, Adv Inst Mfg High Tech Innovat AIM HI, Chiayi, Taiwan
[3] Natl Chung Cheng Univ, Dept Mech Engn, Chiayi 62102, Taiwan
关键词
chaos control; deep reinforcement learning; Lorenz chaotic system; proximal policy optimization;
D O I
10.1002/rnc.6988
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep reinforcement learning (DRL) algorithms are suitable for modeling and controlling complex systems. Methods for controlling chaos, a difficult task, require improvement. In this article, we present a DRL-based control method that can control a nonlinear chaotic system without any prior knowledge of the system's equations. We use proximal policy optimization (PPO) to train an agent. The environment is a Lorenz chaotic system, and our goal is to stabilize this chaotic system as quickly as possible and minimize the error by adding extra control terms to the chaotic system. Therefore, the reward function accounts for the total triaxial error. The experimental results demonstrated that the trained agent can rapidly suppress chaos in the system, regardless of the system's random initial conditions. A comprehensive comparison of different DRL algorithms indicated that PPO is the most efficient and effective algorithm for controlling the chaotic system. Moreover, different maximum control forces were applied to determine the relationship between the control forces and controller performance. To verify the robustness of the controller, random disturbances were introduced during training and testing, and the empirical results indicated that the agent trained with random noise performed better. The chaotic system has highly nonlinear characteristics and is extremely sensitive to initial conditions, and DRL is suitable for modeling such systems.
引用
收藏
页码:586 / 601
页数:16
相关论文
共 44 条
[1]   Convolutional Neural Networks for Speech Recognition [J].
Abdel-Hamid, Ossama ;
Mohamed, Abdel-Rahman ;
Jiang, Hui ;
Deng, Li ;
Penn, Gerald ;
Yu, Dong .
IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2014, 22 (10) :1533-1545
[2]  
Al-Hayali S.Y., 2020, TELKOMNIKA (Telecommunication Computing Electronics and Control), V18, P994, DOI [10.12928/TELKOMNIKA.v18i2.14301, DOI 10.12928/TELKOMNIKA.V18I2.14301]
[3]   Performance improvement of direct torque control for induction motor drive via fuzzy logic-feedback linearization: Simulation and experimental assessment [J].
Ammar, Abdelkarim .
COMPEL-THE INTERNATIONAL JOURNAL FOR COMPUTATION AND MATHEMATICS IN ELECTRICAL AND ELECTRONIC ENGINEERING, 2019, 38 (02) :672-692
[4]  
[Anonymous], PROXIMAL POLICY OPTI
[5]   Deep Reinforcement Learning A brief survey [J].
Arulkumaran, Kai ;
Deisenroth, Marc Peter ;
Brundage, Miles ;
Bharath, Anil Anthony .
IEEE SIGNAL PROCESSING MAGAZINE, 2017, 34 (06) :26-38
[6]  
Brunton S. L., 2019, Data Driven Science Engineering. Machine Learning, Dynamical Systems, DOI [DOI 10.1017/9781108380690, 10.1017/9781108380690]
[7]   Discovering governing equations from data by sparse identification of nonlinear dynamical systems [J].
Brunton, Steven L. ;
Proctor, Joshua L. ;
Kutz, J. Nathan .
PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2016, 113 (15) :3932-3937
[8]   Control of chaotic systems by deep reinforcement learning [J].
Bucci, M. A. ;
Semeraro, O. ;
Allauzen, A. ;
Wisniewski, G. ;
Cordier, L. ;
Mathelin, L. .
PROCEEDINGS OF THE ROYAL SOCIETY A-MATHEMATICAL PHYSICAL AND ENGINEERING SCIENCES, 2019, 475 (2231)
[9]  
Chen CG, 2019, IEEE INT CONF ROBOT, P6015, DOI [10.1109/icra.2019.8794134, 10.1109/ICRA.2019.8794134]
[10]  
Deep LY., ARXIV