Adaptive PI Controller Based on a Reinforcement Learning Algorithm for Speed Control of a DC Motor

被引:7
作者
Alejandro-Sanjines, Ulbio [1 ]
Maisincho-Jivaja, Anthony [1 ]
Asanza, Victor [2 ]
Lorente-Leyva, Leandro L. [2 ,3 ]
Peluffo-Ordonez, Diego H. [2 ,4 ,5 ]
机构
[1] Escuela Super Politecn Litoral, Guayaquil 090903, Ecuador
[2] SDAS Res Grp, Ben Guerir 43150, Morocco
[3] Univ UTE, Fac Law Adm & Social Sci, Quito 170147, Ecuador
[4] Mohammed VI Polytech Univ, Coll Comp, Ben Guerir 47963, Morocco
[5] Corp Univ Autonoma Narino, Fac Engn, Pasto 520001, Colombia
关键词
reinforcement learning; artificial intelligence; adaptive PI; DDPG TD3; neural network; STABILITY; SYSTEMS;
D O I
10.3390/biomimetics8050434
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
Automated industrial processes require a controller to obtain an output signal similar to the reference indicated by the user. There are controllers such as PIDs, which are efficient if the system does not change its initial conditions. However, if this is not the case, the controller must be retuned, affecting production times. In this work, an adaptive PID controller is developed for a DC motor speed plant using an artificial intelligence algorithm based on reinforcement learning. This algorithm uses an actor-critic agent, where its objective is to optimize the actor's policy and train a critic for rewards. This will generate the appropriate gains without the need to know the system. The Deep Deterministic Policy Gradient with Twin Delayed (DDPG TD3) was used, with a network composed of 300 neurons for the agent's learning. Finally, the performance of the obtained controller is compared with a classical control one using a cost function.
引用
收藏
页数:26
相关论文
共 69 条
[1]  
Al-Bargothi SN, 2019, Indonesian Journal of Electrical Engineering and Computer Science, V16, P1221, DOI [10.11591/ijeecs.v16.i3.pp1221-1228, 10.11591/ijeecs.v16.i3, DOI 10.11591/IJEECS.V16.I3]
[2]   Optimal Coordinated Control Strategy of Clustered DC Microgrids under Load-Generation Uncertainties Based on GWO [J].
Al-Tameemi, Zaid Hamid Abdulabbas ;
Lie, Tek Tjing ;
Foo, Gilbert ;
Blaabjerg, Frede .
ELECTRONICS, 2022, 11 (08)
[3]   Control Barrier Function Based Quadratic Programs for Safety Critical Systems [J].
Ames, Aaron D. ;
Xu, Xiangru ;
Grizzle, Jessy W. ;
Tabuada, Paulo .
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2017, 62 (08) :3861-3876
[4]  
Anugula R., 2021, P 2021 IEEE INT C TE, P1, DOI [10.1109/TRIBES52498.2021.9751630, DOI 10.1109/TRIBES52498.2021.9751630]
[5]  
Berry MichaelW., 2019, Supervised and Unsupervised Learning for Data Science
[6]  
Bhaumik A, 2016, PROCEEDINGS OF IEEE INTERNATIONAL CONFERENCE ON CIRCUIT, POWER AND COMPUTING TECHNOLOGIES (ICCPCT 2016)
[7]  
Bošković M, 2019, IJEEC - INTERNATIONAL JOURNAL OF ELECTRICAL ENGINEERING AND COMPUTING, V2, DOI [10.7251/ijeec1801001b, 10.7251/IJEEC1801001B, DOI 10.7251/IJEEC1801001B]
[8]  
Brunton SL, 2019, DATA-DRIVEN SCIENCE AND ENGINEERING: MACHINE LEARNING, DYNAMICAL SYSTEMS, AND CONTROL, P1, DOI 10.1017/9781108380690
[9]  
Cantarero T.A., 2015, Diseno del Controlador PID
[10]   An adaptive deep reinforcement learning approach for MIMO PID control of mobile robots [J].
Carlucho, Ignacio ;
De Paula, Mariano ;
Acosta, Gerardo G. .
ISA TRANSACTIONS, 2020, 102 :280-294