Using Q-learning and genetic algorithms to improve the efficiency of weight adjustments for optimal control and design problems

被引:7
|
作者
Kamali, Kaivan [1 ]
Jiang, L. J. [2 ]
Yen, John [1 ]
Wang, K. W. [2 ]
机构
[1] Penn State Univ, Coll Informat Sci & Technol, Lab Intelligent Agents, University Pk, PA 16802 USA
[2] Penn State Univ, Dept Mech & Nucl Engn, Struct Dynam & Control Lab, University Pk, PA 16802 USA
关键词
optimal control; weight selection; Q-learning; genetic algorithms;
D O I
10.1115/1.2739502
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
In traditional optimal control and design problems, the control gains and design parameters are usually derived to minimize a cost function reflecting the system performance and control effort. One major challenge of such approaches is the selection of weighting matrices in the cost function, which are usually determined via trial-and-error and human intuition. While various techniques have been proposed to automate the weight selection process, they either can not address complex design problems or suffer from slow convergence rate and high computational costs. We propose a layered approach based on Q-learning, a reinforcement learning technique, on top of genetic algorithms (GA) to determine the best weightings for optimal control and design problems. The layered approach allows for reuse of knowledge. Knowledge obtained via Q-learning in a design problem can be used to speed up the convergence rate of a similar design problem. Moreover the layered approach allows for solving optimizations that cannot be solved by GA alone. To test the proposed method, we perform numerical experiments on a sample active-passive hybrid vibration control problem, namely adaptive structures with active-passive hybrid piezoelectric networks. These numerical experiments show that the proposed Q-learning scheme is a promising approach for automation of weight selection for complex design problems.
引用
收藏
页码:302 / 308
页数:7
相关论文
共 50 条
  • [1] On the effect of probing noise in optimal control LQR via Q-learning using adaptive filtering algorithms
    Lopez Yanez, Williams Jesus
    de Souza, Francisco das Chagas
    EUROPEAN JOURNAL OF CONTROL, 2022, 65
  • [2] Weight Adjustments in a Routing Algorithm for Wireless Sensor and Actuator Networks Using Q-Learning
    Kuenzel, Gustavo
    Cainelli, Gustavo Pedroso
    Mueller, Ivan
    Pereira, Carlos Eduardo
    IFAC PAPERSONLINE, 2018, 51 (10): : 58 - 63
  • [3] Input-Decoupled Q-Learning for Optimal Control
    Minh Q. Phan
    Seyed Mahdi B. Azad
    The Journal of the Astronautical Sciences, 2020, 67 : 630 - 656
  • [4] Input-Decoupled Q-Learning for Optimal Control
    Phan, Minh Q.
    Azad, Seyed Mahdi B.
    JOURNAL OF THE ASTRONAUTICAL SCIENCES, 2020, 67 (02): : 630 - 656
  • [5] A Combined Policy Gradient and Q-learning Method for Data-driven Optimal Control Problems
    Lin, Mingduo
    Liu, Derong
    Zhao, Bo
    Dai, Qionghai
    Dong, Yi
    2019 9TH INTERNATIONAL CONFERENCE ON INFORMATION SCIENCE AND TECHNOLOGY (ICIST2019), 2019, : 6 - 10
  • [6] Output Feedback Optimal Tracking Control Using Reinforcement Q-Learning
    Rizvi, Syed Ali Asad
    Lin, Zongli
    2018 ANNUAL AMERICAN CONTROL CONFERENCE (ACC), 2018, : 3423 - 3428
  • [7] On-policy Q-learning for Adaptive Optimal Control
    Jha, Sumit Kumar
    Bhasin, Shubhendu
    2014 IEEE Symposium on Adaptive Dynamic Programming and Reinforcement Learning (ADPRL), 2014, : 301 - 306
  • [8] Optimizing traffic flow with Q-learning and genetic algorithm for congestion control
    Deepika, Gitanjali
    Pandove, Gitanjali
    EVOLUTIONARY INTELLIGENCE, 2024, 17 (5-6) : 4179 - 4197
  • [9] Q-learning and policy iteration algorithms for stochastic shortest path problems
    Huizhen Yu
    Dimitri P. Bertsekas
    Annals of Operations Research, 2013, 208 : 95 - 132
  • [10] Q-learning and policy iteration algorithms for stochastic shortest path problems
    Yu, Huizhen
    Bertsekas, Dimitri P.
    ANNALS OF OPERATIONS RESEARCH, 2013, 208 (01) : 95 - 132