Hierarchical Interpretable Imitation Learning for End-to-End Autonomous Driving

被引:67
|
作者
Teng, Siyu [1 ,2 ]
Chen, Long [3 ,4 ]
Ai, Yunfeng [5 ]
Zhou, Yuanye [6 ]
Xuanyuan, Zhe [1 ]
Hu, Xuemin [7 ]
机构
[1] HKBU United Int Coll, BNU, Zhuhai 999077, Peoples R China
[2] Hong Kong Baptist Univ, Kowloon, Hong Kong 999077, Peoples R China
[3] Chinese Acad Sci, Inst Automat, State Key Lab Management & Control Complex Syst, Beijing 100190, Peoples R China
[4] Waytous Inc Qingdao, Qingdao 266109, Peoples R China
[5] Univ Chinese Acad Sci, Beijing 100049, Peoples R China
[6] Malardalen Univ, S-72214 Vasteras, Sweden
[7] Hubei Univ, Sch Comp Sci & Informat Engn, Wuhan 430062, Peoples R China
来源
IEEE TRANSACTIONS ON INTELLIGENT VEHICLES | 2023年 / 8卷 / 01期
基金
中国国家自然科学基金;
关键词
Semantics; Data models; Autonomous vehicles; Cameras; Reinforcement learning; Predictive models; Robustness; Autonomous driving; imitation learning; motion planning; end-to-End driving; interpretability;
D O I
10.1109/TIV.2022.3225340
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
End-to-end autonomous driving provides a simple and efficient framework for autonomous driving systems, which can directly obtain control commands from raw perception data. However, it fails to address stability and interpretability problems in complex urban scenarios. In this paper, we construct a two-stage end-to-end autonomous driving model for complex urban scenarios, named HIIL (Hierarchical Interpretable Imitation Learning), which integrates interpretable BEV mask and steering angle to solve the problems shown above. In Stage One, we propose a pretrained Bird's Eye View (BEV) model which leverages a BEV mask to present an interpretation of the surrounding environment. In Stage Two, we construct an Interpretable Imitation Learning (IIL) model that fuses BEV latent feature from Stage One with an additional steering angle from Pure-Pursuit (PP) algorithm. In the HIIL model, visual information is converted to semantic images by the semantic segmentation network, and the semantic images are encoded to extract the BEV latent feature, which are decoded to predict BEV masks and fed to the IIL as perception data. In this way, the BEV latent feature bridges the BEV and IIL models. Visual information can be supplemented by the calculated steering angle for PP algorithm, speed vector, and location information, thus it could have better performance in complex and terrible scenarios. Our HIIL model meets an urgent requirement for interpretability and robustness of autonomous driving. We validate the proposed model in the CARLA simulator with extensive experiments which show remarkable interpretability, generalization, and robustness capability in unknown scenarios for navigation tasks.
引用
收藏
页码:673 / 683
页数:11
相关论文
共 50 条
  • [31] End-to-End Autonomous Driving With Semantic Depth Cloud Mapping and Multi-Agent
    Natan, Oskar
    Miura, Jun
    IEEE TRANSACTIONS ON INTELLIGENT VEHICLES, 2023, 8 (01): : 557 - 571
  • [32] End-to-end Spatiotemporal Attention Model for Autonomous Driving
    Zhao, Ruijie
    Zhang, Yanxin
    Huang, Zhiqing
    Yin, Chenkun
    PROCEEDINGS OF 2020 IEEE 4TH INFORMATION TECHNOLOGY, NETWORKING, ELECTRONIC AND AUTOMATION CONTROL CONFERENCE (ITNEC 2020), 2020, : 2649 - 2653
  • [33] SEECAD: Semantic End-to-End Communication for Autonomous Driving
    Ribouh, Soheyb
    Hadid, Abdenour
    2024 35TH IEEE INTELLIGENT VEHICLES SYMPOSIUM, IEEE IV 2024, 2024, : 1808 - 1813
  • [34] XBG: End-to-End Imitation Learning for Autonomous Behaviour in Human-Robot Interaction and Collaboration
    Cardenas-Perez, Carlos
    Romualdi, Giulio
    Elobaid, Mohamed
    Dafarra, Stefano
    L'Erario, Giuseppe
    Traversaro, Silvio
    Morerio, Pietro
    Del Bue, Alessio
    Pucci, Daniele
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2024, 9 (12): : 11617 - 11624
  • [35] Multi-modal policy fusion for end-to-end autonomous driving
    Huang, Zhenbo
    Sun, Shiliang
    Zhao, Jing
    Mao, Liang
    INFORMATION FUSION, 2023, 98
  • [36] An End-to-End Motion Planner Using Sensor Fusion for Autonomous Driving
    Thu, Nguyen Thi Hoai
    Han, Dong Seog
    2023 INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE IN INFORMATION AND COMMUNICATION, ICAIIC, 2023, : 678 - 683
  • [37] End-to-End Learning with Memory Models for Complex Autonomous Driving Tasks in Indoor Environments
    Zhihui Lai
    Thomas Bräunl
    Journal of Intelligent & Robotic Systems, 2023, 107
  • [38] End-to-End Learning with Memory Models for Complex Autonomous Driving Tasks in Indoor Environments
    Lai, Zhihui
    Braunl, Thomas
    JOURNAL OF INTELLIGENT & ROBOTIC SYSTEMS, 2023, 107 (03)
  • [39] Minimizing Probabilistic End-to-end Latencies of Autonomous Driving Systems
    Han, Taeho
    Kim, Kanghee
    2023 IEEE 29TH REAL-TIME AND EMBEDDED TECHNOLOGY AND APPLICATIONS SYMPOSIUM, RTAS, 2023, : 27 - 39
  • [40] An End-to-End solution to Autonomous Driving based on Xilinx FPGAd
    Wu, Tianze
    Liu, Weiyi
    Jin, Yongwei
    2019 INTERNATIONAL CONFERENCE ON FIELD-PROGRAMMABLE TECHNOLOGY (ICFPT 2019), 2019, : 427 - 430