Deep reinforcement learning with predictive auxiliary task for autonomous train collision avoidance

被引:0
作者
Plissonneau, Antoine [1 ,2 ]
Jourdan, Luca [1 ]
Trentesaux, Damien [2 ]
Abdi, Lotfi [1 ]
Sallak, Mohamed [1 ,3 ]
Bekrar, Abdelghani [2 ]
Quost, Benjamin [1 ,3 ]
Schoen, Walter [1 ,3 ]
机构
[1] Railenium, Valenciennes, France
[2] Univ Polytech Hauts De France, CNRS, LAMIH, UMR 8201, F-59313 Valenciennes, France
[3] Univ Technol Compiegne, CNRS, Heudiasyc Heurist & Diagnost Syst Complexes, CS 60 319, F-60203 Compiegne, France
关键词
Autonomous train; Collision avoidance; Deep reinforcement learning; Auxiliary task; Interpretability; NEURAL-NETWORKS; NAVIGATION; LEVEL;
D O I
10.1016/j.jrtpm.2024.100453
中图分类号
U [交通运输];
学科分类号
08 ; 0823 ;
摘要
The contribution of this paper consists of a deep reinforcement learning (DRL) based method for autonomous train collision avoidance. While DRL applied to autonomous vehicles' collision avoidance has shown interesting results compared to traditional methods, train -like vehicles are not currently covered. In addition, DRL applied to collision avoidance suffers from sparse rewards, which can lead to poor convergence and long training time. To overcome these limitations, this paper proposes a method for training a reinforcement learning (RL) agent for collision avoidance using local obstacle information mapped into occupancy grids. This method also integrates a network architecture containing a predictive auxiliary task consisting in future state prediction and encouraging the intermediate representation to be predictive of obstacle trajectories. A comparison study conducted on multiple simulated scenarios demonstrates that the trained policy outperforms other deep-learning-based policies as well as human driving in terms of both safety and efficiency. As a first step toward the certification of a DRL based method, this paper proposes to approximate the policy learned by the RL agent with an interpretable decision tree. Although this approximation results in a loss of performance, it enables a safety analysis of the learned function and thus paves the way to use the strengths of RL in certifiable algorithms. As this work is pioneering the use of RL for collision avoidance of rail-guided vehicles, and to facilitate future work by other engineers and researchers, a RL-ready simulator is provided with this paper.
引用
收藏
页数:20
相关论文
共 50 条
[21]   A novel intelligent collision avoidance algorithm based on deep reinforcement learning approach for USV [J].
Fan, Yunsheng ;
Sun, Zhe ;
Wang, Guofeng .
OCEAN ENGINEERING, 2023, 287
[22]   Research on Collision Avoidance Algorithm of Unmanned Surface Vehicle Based on Deep Reinforcement Learning [J].
Xia, Jiawei ;
Zhu, Xufang ;
Liu, Zhikun ;
Luo, Yasong ;
Wu, Zhaodong ;
Wu, Qiuhan .
IEEE SENSORS JOURNAL, 2023, 23 (11) :11262-11273
[23]   Deep-Reinforcement-Learning-Based Collision Avoidance in UAV Environment [J].
Ouahouah, Sihem ;
Bagaa, Miloud ;
Prados-Garzon, Jonathan ;
Taleb, Tarik .
IEEE INTERNET OF THINGS JOURNAL, 2022, 9 (06) :4015-4030
[24]   Collision avoidance for an unmanned surface vehicle using deep reinforcement learning [J].
Woo, Joohyun ;
Kim, Nakwan .
OCEAN ENGINEERING, 2020, 199
[25]   Multigoal Visual Navigation With Collision Avoidance via Deep Reinforcement Learning [J].
Xiao, Wendong ;
Yuan, Liang ;
He, Li ;
Ran, Teng ;
Zhang, Jianbo ;
Cui, Jianping .
IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT, 2022, 71
[26]   Collision Avoidance in Pedestrian-Rich Environments With Deep Reinforcement Learning [J].
Everett, Michael ;
Chen, Yu Fan ;
How, Jonathan P. .
IEEE ACCESS, 2021, 9 :10357-10377
[27]   Space Manipulator Collision Avoidance Using a Deep Reinforcement Learning Control [J].
Blaise, James ;
Bazzocchi, Michael C. F. .
AEROSPACE, 2023, 10 (09)
[28]   Vision-guided Collision Avoidance through Deep Reinforcement Learning [J].
Song, Sirui ;
Zhang, Yuanhang ;
Qin, Xi ;
Saunders, Kirk ;
Liu, Jundong .
PROCEEDINGS OF THE 2021 IEEE NATIONAL AEROSPACE AND ELECTRONICS CONFERENCE (NAECON), 2021, :191-194
[29]   COLREGs-abiding hybrid collision avoidance algorithm based on deep reinforcement learning for USVs [J].
Xu, Xinli ;
Lu, Yu ;
Liu, Gang ;
Cai, Peng ;
Zhang, Weidong .
OCEAN ENGINEERING, 2022, 247
[30]   Collaborative Collision Avoidance Approach for USVs Based on Multi-Agent Deep Reinforcement Learning [J].
Wang, Zhiwen ;
Chen, Pengfei ;
Chen, Linying ;
Mou, Junmin .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2025, 26 (04) :4780-4794