Path planning via reinforcement learning with closed-loop motion control and field tests

被引:0
|
作者
Feher, Arpad [1 ]
Domina, Adam [2 ]
Bardos, Adam [2 ]
Aradi, Szilard [1 ]
Becsi, Tamas [1 ]
机构
[1] Budapest Univ Technol & Econ, Fac Transportat Engn & Vehicle Engn, Dept Control Transportat & Vehicle Syst, Muegyet Rkp 3, H-1111 Budapest, Hungary
[2] Budapest Univ Technol & Econ, Dept Automot Technol, Fac Transportat Engn & Vehicle Engn, Muegyetem Rkp 3, H-1111 Budapest, Hungary
关键词
Vehicle dynamics; Advanced driver assistance systems; Machine learning; Reinforcement learning; Model predictive control; ACTIVE STEERING CONTROL; MODEL; SIMULATION; VEHICLES;
D O I
10.1016/j.engappai.2024.109870
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Performing evasive maneuvers with highly automated vehicles is a challenging task. The algorithm must fulfill safety constraints and complete the task while keeping the car in a controllable state. Furthermore, considering all aspects of vehicle dynamics, the path generation problem is numerically complex. Hence its classical solutions can hardly meet real-time requirements. On the other hand, single reinforcement learning based approaches only could handle this problem as a simple driving task and would not provide feasibility information on the whole task's horizon. Therefore, this paper presents a hierarchical method for obstacle avoidance of an automated vehicle to overcome this issue, where the geometric path generation is provided by a single-step continuous Reinforcement Learning agent, while a model-predictive controller deals with lateral control to perform a double lane change maneuver. As the agent plays the optimization role in this architecture, it is trained in various scenarios to provide the necessary parameters fora geometric path generator in a onestep neural network output. During the training, the controller that follows the track evaluates the feasibility of the generated path whose performance metrics provide feedback to the agent so it can further improve its performance. The framework can train an agent fora given problem with various parameters. Asa use case, it is presented as a static obstacle avoidance maneuver. the proposed framework was tested on an automotive proving ground with the geometric constraints of the ISO-3888-2 test. The results proved its real-time capability and performance compared to human drivers' abilities.
引用
收藏
页数:13
相关论文
共 50 条
  • [41] Energy-Efficient Motion Planning and Control for Robotic Arms via Deep Reinforcement Learning
    Shen, Tan
    Liu, Xing
    Dong, Yunlong
    Yuan, Ye
    2022 34TH CHINESE CONTROL AND DECISION CONFERENCE, CCDC, 2022, : 5502 - 5507
  • [42] Behavioral analysis of differential hebbian learning in closed-loop systems
    Kulvicius, Tomas
    Kolodziejski, Christoph
    Tamosiunaite, Minija
    Porr, Bernd
    Woergoetter, Florentin
    BIOLOGICAL CYBERNETICS, 2010, 103 (04) : 255 - 271
  • [43] Closed-Loop Planning for Disaster Evacuation with Stochastic Arrivals
    Sidrane, Chelsea
    Kochenderfer, Mykel J.
    2018 21ST INTERNATIONAL CONFERENCE ON INTELLIGENT TRANSPORTATION SYSTEMS (ITSC), 2018, : 2544 - 2549
  • [44] The Intelligent Path Planning System of Agricultural Robot via Reinforcement Learning
    Yang, Jiachen
    Ni, Jingfei
    Li, Yang
    Wen, Jiabao
    Chen, Desheng
    SENSORS, 2022, 22 (12)
  • [45] Closed-Loop Reinforcement Learning Based Deep Brain Stimulation Using SpikerNet: A Computational Model
    Coventry, Brandon S.
    Bartlett, Edward L.
    2023 11TH INTERNATIONAL IEEE/EMBS CONFERENCE ON NEURAL ENGINEERING, NER, 2023,
  • [46] Tests on a virtual patient for an observer-based, closed-loop control of plasma glycemia
    Palumbo, P.
    Pizzichelli, G.
    Panunzi, S.
    Pepe, P.
    De Gaetano, A.
    2011 50TH IEEE CONFERENCE ON DECISION AND CONTROL AND EUROPEAN CONTROL CONFERENCE (CDC-ECC), 2011, : 6936 - 6941
  • [47] Learning Realistic Traffic Agents in Closed-loop
    Zhang, Chris
    Tu, James
    Zhang, Lunjun
    Wong, Kelvin
    Suo, Simon
    Urtasun, Raquel
    CONFERENCE ON ROBOT LEARNING, VOL 229, 2023, 229
  • [48] Wind-Tunnel Tests for Active Flutter Control and Closed-Loop Flutter Identification
    Huang, Rui
    Zhao, Yonghui
    Hu, Haiyan
    AIAA JOURNAL, 2016, 54 (07) : 2089 - 2099
  • [49] Reinforcement learning for closed-loop regulation of cardiovascular system with vagus nerve stimulation: a computational study
    Sarikhani, Parisa
    Hsu, Hao-Lun
    Zeydabadinezhad, Mahmoud
    Yao, Yuyu
    Kothare, Mayuresh
    Mahmoudi, Babak
    JOURNAL OF NEURAL ENGINEERING, 2024, 21 (03)
  • [50] Path planning of virtual human by using reinforcement learning
    He, Yue-Sheng
    Tang, Yuan-Yan
    PROCEEDINGS OF 2008 INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND CYBERNETICS, VOLS 1-7, 2008, : 987 - 992