Adaptive Sliding Mode Disturbance Observer and Deep Reinforcement Learning Based Motion Control for Micropositioners

被引:7
作者
Liang, Shiyun [1 ,2 ]
Xi, Ruidong [1 ,2 ]
Xiao, Xiao [3 ]
Yang, Zhixin [1 ,2 ]
机构
[1] Univ Macau, State Key Lab Internet Things Smart City, Macau 999078, Peoples R China
[2] Univ Macau, Dept Electromech Engn, Macau 999078, Peoples R China
[3] Southern Univ Sci & Technol, Dept Elect & Elect Engn, Shenzhen 518055, Peoples R China
关键词
micropositioners; reinforcement learning; disturbance observer; deep deterministic policy gradient; SYSTEMS;
D O I
10.3390/mi13030458
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
The motion control of high-precision electromechanitcal systems, such as micropositioners, is challenging in terms of the inherent high nonlinearity, the sensitivity to external interference, and the complexity of accurate identification of the model parameters. To cope with these problems, this work investigates a disturbance observer-based deep reinforcement learning control strategy to realize high robustness and precise tracking performance. Reinforcement learning has shown great potential as optimal control scheme, however, its application in micropositioning systems is still rare. Therefore, embedded with the integral differential compensator (ID), deep deterministic policy gradient (DDPG) is utilized in this work with the ability to not only decrease the state error but also improve the transient response speed. In addition, an adaptive sliding mode disturbance observer (ASMDO) is proposed to further eliminate the collective effect caused by the lumped disturbances. The micropositioner controlled by the proposed algorithm can track the target path precisely with less than 1 mu m error in simulations and actual experiments, which shows the sterling performance and the accuracy improvement of the controller.
引用
收藏
页数:21
相关论文
共 47 条
[11]   Design of Reconfigurable Planar Micro-Positioning Stages Based on Function Modules [J].
Ding, Bingxiao ;
Yang, Zhi-Xin ;
Xiao, Xiao ;
Zhang, Geng .
IEEE ACCESS, 2019, 7 :15102-15112
[12]   Adaptive Fuzzy Sliding Mode Control for a Micro Gyroscope with Backstepping Controller [J].
Fei, Juntao ;
Fang, Yunmei ;
Yuan, Zhuli .
MICROMACHINES, 2020, 11 (11)
[13]   Fractional delay filter based repetitive control for precision tracking: Design and application to a piezoelectric nanopositioning stage [J].
Feng, Zhao ;
Ming, Min ;
Ling, Jie ;
Xiao, Xiaohui ;
Yang, Zhi-Xin ;
Wan, Feng .
MECHANICAL SYSTEMS AND SIGNAL PROCESSING, 2022, 164
[14]   A PID-Type Fuzzy Logic Controller-Based Approach for Motion Control Applications [J].
Garcia-Martinez, Jose R. ;
Cruz-Miguel, Edson E. ;
Carrillo-Serrano, Roberto, V ;
Mendoza-Mondragon, Fortino ;
Toledano-Ayala, Manuel ;
Rodriguez-Resendiz, Juvenal .
SENSORS, 2020, 20 (18) :1-19
[15]   A New Seven-Segment Profile Algorithm for an Open Source Architecture in a Hybrid Electronic Platform [J].
Garcia-Martinez, Jose R. ;
Rodriguez-Resendiz, Juvenal ;
Cruz-Miguel, Edson E. .
ELECTRONICS, 2019, 8 (06)
[16]   Path tracking control of electromechanical micro-positioner by considering control effort of the system [J].
Gharib, Mohammad Reza ;
Koochi, Ali ;
Ghorbani, Mojtaba .
PROCEEDINGS OF THE INSTITUTION OF MECHANICAL ENGINEERS PART I-JOURNAL OF SYSTEMS AND CONTROL ENGINEERING, 2021, 235 (06) :984-991
[17]   Event-Triggered Reinforcement Learning-Based Adaptive Tracking Control for Completely Unknown Continuous-Time Nonlinear Systems [J].
Guo, Xinxin ;
Yan, Weisheng ;
Cui, Rongxin .
IEEE TRANSACTIONS ON CYBERNETICS, 2020, 50 (07) :3231-3242
[18]   Reinforcement learning control of constrained dynamic systems with uniformly ultimate boundedness stability guarantee [J].
Han, Minghao ;
Tian, Yuan ;
Zhang, Lixian ;
Wang, Jun ;
Pan, Wei .
AUTOMATICA, 2021, 129
[19]  
Hasselt H. V., 2010, Advances in Neural Information Processing Systems, V23, P2613
[20]   Optimal and Autonomous Control Using Reinforcement Learning: A Survey [J].
Kiumarsi, Bahare ;
Vamvoudakis, Kyriakos G. ;
Modares, Hamidreza ;
Lewis, Frank L. .
IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2018, 29 (06) :2042-2062