Reinforcement Learning-based Data-driven Control Design for Motion Control Systems

被引:1
作者
Deng, Zhengqi [1 ]
Huo, Xin [1 ]
Du, Qinlong [1 ]
Liu, Qingquan [1 ]
机构
[1] Harbin Inst Technol, Control & Simulat Ctr, Harbin 150080, Peoples R China
来源
PROCEEDINGS OF THE 36TH CHINESE CONTROL AND DECISION CONFERENCE, CCDC 2024 | 2024年
基金
黑龙江省自然科学基金; 中国国家自然科学基金;
关键词
Motion control systems; Data-driven control; Reinforcement learning (RL); Deep deterministic policy gradient (DDPG); Twin delayed deep deterministic policy gradient (TD3);
D O I
10.1109/CCDC62350.2024.10587783
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Motion control systems are widely used in many fields of industry. Conventional control schemes are highly dependent on the system model to be designed. The performance of design would be greatly reduced, when the system exists unknown disturbances or uncertainty. Therefore, some scholars pointed out that the dependency on the system models can be eliminated by data-driven design schemes. In this paper, the reinforcement learning-based methods are included, which appeal to attentions gradually. The disturbances rejection problem for motion control systems is studied based on reinforcement learning. Considering the continuity of state space and action space, a method based on deep reinforcement learning algorithm is proposed to reject the periodic disturbances. Proposed deep deterministic policy gradient (DDPG) and twin delayed deep deterministic policy gradient (TD3) based algorithms are compared in simulation. The simulation results show that the periodic disturbances of the motion control systems can be rejected effectively with the proposed reinforcement learning controller.
引用
收藏
页码:5745 / 5749
页数:5
相关论文
共 13 条
[1]  
Fujimoto S, 2018, PR MACH LEARN RES, V80
[2]   From model-based control to data-driven control: Survey, classification and perspective [J].
Hou, Zhong-Sheng ;
Wang, Zhuo .
INFORMATION SCIENCES, 2013, 235 :3-35
[3]   Voronoi-Based Multi-Robot Autonomous Exploration in Unknown Environments via Deep Reinforcement Learning [J].
Hu, Junyan ;
Niu, Hanlin ;
Carrasco, Joaquin ;
Lennox, Barry ;
Arvin, Farshad .
IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2020, 69 (12) :14413-14423
[4]   Data-driven Iterative Tuning for Rejecting Spatial Periodic Disturbances Combined [J].
Huo, Xin ;
Wu, Aijing ;
Wang, Ruichao ;
Ma, Kemao .
IFAC PAPERSONLINE, 2020, 53 (02) :3959-3964
[5]   Efficient Load Frequency Control of Renewable Integrated Power System: A Twin Delayed DDPG-Based Deep Reinforcement Learning Approach [J].
Khalid, Junaid ;
Ramli, Makbul A. M. ;
Khan, Muhammad Saud ;
Hidayat, Taufal .
IEEE ACCESS, 2022, 10 :51561-51574
[6]  
Li J., 2020, IEEE ACCESS, V8, P156839
[7]  
Lillicrap Timothy P., 2015, INT C LEARNING REPRE
[8]   On Research Progress and Trend of Motion Contro in Industrial Robot Applications [J].
Luo, Xianxi ;
Xu, Liqin .
2020 CHINESE AUTOMATION CONGRESS (CAC 2020), 2020, :5944-5947
[9]   Applications of Deep Reinforcement Learning in Communications and Networking: A Survey [J].
Luong, Nguyen Cong ;
Hoang, Dinh Thai ;
Gong, Shimin ;
Niyato, Dusit ;
Wang, Ping ;
Liang, Ying-Chang ;
Kim, Dong In .
IEEE COMMUNICATIONS SURVEYS AND TUTORIALS, 2019, 21 (04) :3133-3174
[10]   Human-level control through deep reinforcement learning [J].
Mnih, Volodymyr ;
Kavukcuoglu, Koray ;
Silver, David ;
Rusu, Andrei A. ;
Veness, Joel ;
Bellemare, Marc G. ;
Graves, Alex ;
Riedmiller, Martin ;
Fidjeland, Andreas K. ;
Ostrovski, Georg ;
Petersen, Stig ;
Beattie, Charles ;
Sadik, Amir ;
Antonoglou, Ioannis ;
King, Helen ;
Kumaran, Dharshan ;
Wierstra, Daan ;
Legg, Shane ;
Hassabis, Demis .
NATURE, 2015, 518 (7540) :529-533