Hierarchical residual reinforcement learning based path tracking control method for unmanned bicycle

被引:0
|
作者
Huo, Benyan [1 ]
Yu, Long [1 ]
Liu, Yanhong [1 ]
Chen, Zhang [2 ]
机构
[1] School of Electrical and Information Engineering, Zhengzhou University, Zhengzhou,450001, China
[2] Department of Automation, Tsinghua University, Beijing,100084, China
基金
中国国家自然科学基金;
关键词
Attitude control - Bicycles - Reinforcement learning;
D O I
10.1016/j.robot.2025.104996
中图分类号
学科分类号
摘要
Due to their super under-actuation and high nonlinearity properties, simplification and linearization techniques are necessary in the traditional model-based path tracking controller design of unmanned bicycles, leading to the decrease of their control performance. On the other hand, as one of the most important learning-based methods, deep reinforcement learning (DRL) suffers from low training efficiency and initial safety issues. In this letter, a hierarchical residual reinforcement learning (HRRL)-based path tracking control method is proposed, to address the drawbacks of both traditional and learning-based approaches. The path tracking task is decomposed into two subtasks, i.e., attitude control and position control, and the controllers are designed separately for each subtask. In each controller, a DRL controller is connected to a traditional controller through residual connection. Physical simulation experiments demonstrate that compared to the traditional LQR, LQI, Stanley and DRL approaches, the proposed method can improve the tracking performance of unmanned bicycles and decrease the training time and the tipping number during training. Furthermore, experimental results also show that the proposed controller exhibits a certain level of robustness, enabling effective path tracking in complex terrain. © 2025 Elsevier B.V.
引用
收藏
相关论文
共 50 条
  • [31] Path curvature incorporated reinforcement learning method for accurate path tracking of agricultural vehicles
    Zhang, Linhuan
    Zhang, Ruirui
    Zhang, Danzhu
    Yi, Tongchuan
    Ding, Chenchen
    Chen, Liping
    COMPUTERS AND ELECTRONICS IN AGRICULTURE, 2025, 234
  • [32] A knowledge-guided reinforcement learning method for lateral path tracking
    Hu, Bo
    Zhang, Sunan
    Feng, Yuxiang
    Li, Bingbing
    Sun, Hao
    Chen, Mingyang
    Zhuang, Weichao
    Zhang, Yi
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2025, 139
  • [33] Trajectory Tracking Control of Intelligent Vehicle Based on DDPG Method of Reinforcement Learning
    He, Yi-Lin
    Song, Ruo-Yang
    Ma, Jian
    Zhongguo Gonglu Xuebao/China Journal of Highway and Transport, 2021, 34 (11): : 335 - 348
  • [34] Optimal Hierarchical Learning Path Design With Reinforcement Learning
    Li, Xiao
    Xu, Hanchen
    Zhang, Jinming
    Chang, Hua-hua
    APPLIED PSYCHOLOGICAL MEASUREMENT, 2021, 45 (01) : 54 - 70
  • [35] Surface path tracking method of autonomous surface underwater vehicle based on deep reinforcement learning
    Song, Dalei
    Gan, Wenhao
    Yao, Peng
    Zang, Wenchuan
    Qu, Xiuqing
    NEURAL COMPUTING & APPLICATIONS, 2023, 35 (08): : 6225 - 6245
  • [36] Surface path tracking method of autonomous surface underwater vehicle based on deep reinforcement learning
    Dalei Song
    Wenhao Gan
    Peng Yao
    Wenchuan Zang
    Xiuqing Qu
    Neural Computing and Applications, 2023, 35 : 6225 - 6245
  • [37] Adaptive preview based control system for unmanned vehicle path tracking
    Zhou, Wei
    International Journal of Vehicle Structures and Systems, 2021, 13 (04): : 398 - 404
  • [38] Research on Path Tracking of Unmanned Spray Based on Dual Control Strategy
    Wen, Haojun
    Ma, Xiaodong
    Qin, Chenjian
    Chen, Hao
    Kang, Huanyu
    AGRICULTURE-BASEL, 2024, 14 (04):
  • [39] An Autonomous Path Planning Model for Unmanned Ships Based on Deep Reinforcement Learning
    Guo, Siyu
    Zhang, Xiuguo
    Zheng, Yisong
    Du, Yiquan
    SENSORS, 2020, 20 (02)
  • [40] Path Planning Technology of Unmanned Vehicle Based on Improved Deep Reinforcement Learning
    Zhang, Kai
    Wang, Guile
    Hu, Jinwen
    Xu, Zhao
    Guo, Chubing
    2021 PROCEEDINGS OF THE 40TH CHINESE CONTROL CONFERENCE (CCC), 2021, : 8392 - 8397