Using Q-learning and genetic algorithms to improve the efficiency of weight adjustments for optimal control and design problems

被引:7
作者
Kamali, Kaivan [1 ]
Jiang, L. J. [2 ]
Yen, John [1 ]
Wang, K. W. [2 ]
机构
[1] Penn State Univ, Coll Informat Sci & Technol, Lab Intelligent Agents, University Pk, PA 16802 USA
[2] Penn State Univ, Dept Mech & Nucl Engn, Struct Dynam & Control Lab, University Pk, PA 16802 USA
关键词
optimal control; weight selection; Q-learning; genetic algorithms;
D O I
10.1115/1.2739502
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
In traditional optimal control and design problems, the control gains and design parameters are usually derived to minimize a cost function reflecting the system performance and control effort. One major challenge of such approaches is the selection of weighting matrices in the cost function, which are usually determined via trial-and-error and human intuition. While various techniques have been proposed to automate the weight selection process, they either can not address complex design problems or suffer from slow convergence rate and high computational costs. We propose a layered approach based on Q-learning, a reinforcement learning technique, on top of genetic algorithms (GA) to determine the best weightings for optimal control and design problems. The layered approach allows for reuse of knowledge. Knowledge obtained via Q-learning in a design problem can be used to speed up the convergence rate of a similar design problem. Moreover the layered approach allows for solving optimizations that cannot be solved by GA alone. To test the proposed method, we perform numerical experiments on a sample active-passive hybrid vibration control problem, namely adaptive structures with active-passive hybrid piezoelectric networks. These numerical experiments show that the proposed Q-learning scheme is a promising approach for automation of weight selection for complex design problems.
引用
收藏
页码:302 / 308
页数:7
相关论文
共 50 条
[41]   Basis functions and Genetic Algorithms in norm-optimal Iterative Learning Control [J].
Hatzikos, V ;
Hätönen, J ;
Owens, DH .
INTELLIGENT CONTROL SYSTEMS AND SIGNAL PROCESSING 2003, 2003, :285-290
[42]   Optimal Control for Chemical Reactors with Distributed Parameters Using Genetic Algorithms [J].
Woinaroschy, Alexandru .
CHEMICAL ENGINEERING & TECHNOLOGY, 2019, 42 (11) :2393-2400
[43]   OPTIMAL DESIGN FOR INDUCTION BEATING USING GENETIC ALGORITHMS [J].
Leuca, Teodor ;
Nagy, Stefan ;
Trip, Nistor Daniel ;
Silaghi, Helga ;
Mich-Vancea, Claudiu .
REVUE ROUMAINE DES SCIENCES TECHNIQUES-SERIE ELECTROTECHNIQUE ET ENERGETIQUE, 2015, 60 (02) :133-142
[44]   Solving the optimal path planning of a mobile robot using improved Q-learning [J].
Low, Ee Soong ;
Ong, Pauline ;
Cheah, Kah Chun .
ROBOTICS AND AUTONOMOUS SYSTEMS, 2019, 115 :143-161
[45]   Optimal City Navigation for Pedestrians using Agent-specific Q-learning [J].
Miao, Lei .
2018 ANNUAL AMERICAN CONTROL CONFERENCE (ACC), 2018, :4123-4128
[46]   Optimal distribution network design using genetic algorithms [J].
Türkay, B ;
Artaç, T .
ELECTRIC POWER COMPONENTS AND SYSTEMS, 2005, 33 (05) :513-524
[47]   The Adaptive Optimal Output Feedback Tracking Control of Unknown Discrete-Time Linear Systems Using a Multistep Q-Learning Approach [J].
Dong, Xunde ;
Lin, Yuxin ;
Suo, Xudong ;
Wang, Xihao ;
Sun, Weijie .
MATHEMATICS, 2024, 12 (04)
[48]   Data-Driven Optimal Controller Design for Maglev Train: Q-Learning Method [J].
Xin, Liang ;
Jiang, Hongwei ;
Wen, Tao ;
Long, Zhiqiang .
2022 34TH CHINESE CONTROL AND DECISION CONFERENCE, CCDC, 2022, :1289-1294
[49]   Robust Attitude Control of an Agile Aircraft Using Improved Q-Learning [J].
Zahmatkesh, Mohsen ;
Emami, Seyyed Ali ;
Banazadeh, Afshin ;
Castaldi, Paolo .
ACTUATORS, 2022, 11 (12)
[50]   Hierarchical control of traffic signals using Q-learning with tile coding [J].
Monireh Abdoos ;
Nasser Mozayani ;
Ana L. C. Bazzan .
Applied Intelligence, 2014, 40 :201-213