Path planning via reinforcement learning with closed-loop motion control and field tests

被引:0
|
作者
Feher, Arpad [1 ]
Domina, Adam [2 ]
Bardos, Adam [2 ]
Aradi, Szilard [1 ]
Becsi, Tamas [1 ]
机构
[1] Budapest Univ Technol & Econ, Fac Transportat Engn & Vehicle Engn, Dept Control Transportat & Vehicle Syst, Muegyet Rkp 3, H-1111 Budapest, Hungary
[2] Budapest Univ Technol & Econ, Dept Automot Technol, Fac Transportat Engn & Vehicle Engn, Muegyetem Rkp 3, H-1111 Budapest, Hungary
关键词
Vehicle dynamics; Advanced driver assistance systems; Machine learning; Reinforcement learning; Model predictive control; ACTIVE STEERING CONTROL; MODEL; SIMULATION; VEHICLES;
D O I
10.1016/j.engappai.2024.109870
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Performing evasive maneuvers with highly automated vehicles is a challenging task. The algorithm must fulfill safety constraints and complete the task while keeping the car in a controllable state. Furthermore, considering all aspects of vehicle dynamics, the path generation problem is numerically complex. Hence its classical solutions can hardly meet real-time requirements. On the other hand, single reinforcement learning based approaches only could handle this problem as a simple driving task and would not provide feasibility information on the whole task's horizon. Therefore, this paper presents a hierarchical method for obstacle avoidance of an automated vehicle to overcome this issue, where the geometric path generation is provided by a single-step continuous Reinforcement Learning agent, while a model-predictive controller deals with lateral control to perform a double lane change maneuver. As the agent plays the optimization role in this architecture, it is trained in various scenarios to provide the necessary parameters fora geometric path generator in a onestep neural network output. During the training, the controller that follows the track evaluates the feasibility of the generated path whose performance metrics provide feedback to the agent so it can further improve its performance. The framework can train an agent fora given problem with various parameters. Asa use case, it is presented as a static obstacle avoidance maneuver. the proposed framework was tested on an automotive proving ground with the geometric constraints of the ISO-3888-2 test. The results proved its real-time capability and performance compared to human drivers' abilities.
引用
收藏
页数:13
相关论文
共 50 条
  • [21] Reinforcement Learning for Closed-Loop Propofol Anesthesia: A Study in Human Volunteers
    Moore, Brett L.
    Pyeatt, Larry D.
    Kulkarni, Vivekanand
    Panousis, Periklis
    Padrez, Kevin
    Doufaas, Anthony G.
    JOURNAL OF MACHINE LEARNING RESEARCH, 2014, 15 : 655 - 696
  • [22] Closed-Loop Scheduling and Control for Precision Irrigation
    Nahar, Jannatun
    Liu, Su
    Mao, Yawen
    Liu, Jinfeng
    Shah, Sirish L.
    INDUSTRIAL & ENGINEERING CHEMISTRY RESEARCH, 2019, 58 (26) : 11485 - 11497
  • [23] Conceptual closed-loop design of automotive cooling systems leveraging Reinforcement Learning
    Vanhuyse, Johan
    Bertheaume, Clement
    Gumussoy, Suat
    Nicolai, Mike
    FORSCHUNG IM INGENIEURWESEN-ENGINEERING RESEARCH, 2025, 89 (01):
  • [24] A closed-loop algorithm to detect human face using color and reinforcement learning
    吴东晖
    叶秀清
    顾伟康
    "JournalofZhejiangUniversityScienceJ", 2002, (01) : 73 - 77
  • [25] A closed-loop algorithm to detect human face using color and reinforcement learning
    Wu Dong-hui
    Ye Xiu-qing
    Gu Wei-kang
    Journal of Zhejiang University-SCIENCE A, 2002, 3 (1): : 72 - 76
  • [26] Improving ceramic additive manufacturing via machine learning-enabled closed-loop control
    Zhang, Zhaolong
    Yang, Zhaotong
    Sisson, Richard D.
    Liang, Jianyu
    INTERNATIONAL JOURNAL OF APPLIED CERAMIC TECHNOLOGY, 2022, 19 (02) : 957 - 967
  • [27] Path Planning and Energy Optimization in Optimal Control of Autonomous Wheel Loaders Using Reinforcement Learning
    Sardarmehni, Tohid
    Song, Xingyong
    IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2023, 72 (08) : 9821 - 9834
  • [28] Model-free Reinforcement Learning with a Non-linear Reconstructor for closed-loop Adaptive Optics control with a pyramid wavefront sensor
    Pou, B.
    Smith, J.
    Quinones, E.
    Martin, M.
    Gratadour, D.
    ADAPTIVE OPTICS SYSTEMS VIII, 2022, 12185
  • [29] Model-free closed-loop wind farm control using reinforcement learning with recursive least squares
    Liew, Jaime
    Gocmen, Tuhfe
    Lio, Wai Hou
    Larsen, Gunner Chr.
    WIND ENERGY, 2024, 27 (11) : 1173 - 1187
  • [30] UAV swarm path planning with reinforcement learning for field prospecting
    Puente-Castro, Alejandro
    Rivero, Daniel
    Pazos, Alejandro
    Fernandez-Blanco, Enrique
    APPLIED INTELLIGENCE, 2022, 52 (12) : 14101 - 14118