Optimal combination of imitation and reinforcement learning for self-driving cars

被引:2
|
作者
Youssef F. [1 ]
Houda B. [1 ]
机构
[1] National School of Computer Science and Systems Analysis (ENSIAS), Mohammed V University, Rabat
来源
Revue d'Intelligence Artificielle | 2019年 / 33卷 / 04期
关键词
Behavioral cloning; Deep reinforcement learning; Expert's trust margin; Prioritized experience replay; Simulation environment; Supervised imitation learning;
D O I
10.18280/ria.330402
中图分类号
学科分类号
摘要
The two steps in human intelligence development, namely, mimicking and tentative application of expertise, are reflected by imitation learning (IL) and reinforcement learning (RL) in artificial intelligence (AI). However, the RL process does not always improve the skills learned from expert demonstrations and enhance the algorithm performance. To solve the problem, this paper puts forward a novel algorithm called optimal combination of imitation and reinforcement learning (OCIRL). First, the concept of deep q-learning from demonstrations (DQfD) was introduced to the actor-critic (A2C) model, creating the A2CfD model. Then, a threshold was estimated from a trained IL model with the same inputs and reward function with the DOfD, and applied to the A2CfD model. The threshold represents the minimum reward that conserves the learned expertise. The resulting A2CfDoC model was trained and tested on self-driving cars in both discrete and continuous environments. The results show that the model outperformed several existing algorithms in terms of speed and accuracy. © 2019 Lavoisier. All rights reserved.
引用
收藏
页码:265 / 273
页数:8
相关论文
共 50 条
  • [41] Self-driving cars - the human side
    Szikora, Peter
    Madarasz, Nikolett
    2017 IEEE 14TH INTERNATIONAL SCIENTIFIC CONFERENCE ON INFORMATICS, 2017, : 383 - 387
  • [42] Kangaroos confuse self-driving cars
    Matthews, Jason
    Caudell, Joe N.
    HUMAN-WILDLIFE INTERACTIONS, 2017, 11 (03): : 243 - 244
  • [43] Self-driving cars on the starting grid
    Kowal, Jan
    NEW SCIENTIST, 2016, 229 (3060) : 52 - 52
  • [44] Moral dilemmas in self-driving cars
    Lucifora, Chiara
    Grasso, Giorgio Mario
    Perconti, Pietro
    Plebe, Alessio
    RIVISTA INTERNAZIONALE DI FILOSOFIA E PSICOLOGIA, 2020, 11 (02) : 238 - 250
  • [45] A Combination of Feedback Control and Vision-Based Deep Learning Mechanism for Guiding Self-Driving Cars
    Lin, Wen-Yen
    Hsu, Wang-Hsin
    Chiang, Yi-Yuan
    2018 IEEE INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND VIRTUAL REALITY (AIVR), 2018, : 262 - 266
  • [46] Deep Learning for Self-Driving Cars: Chances and Challenges Extended Abstract
    Rao, Qing
    Frtunikj, Jelena
    PROCEEDINGS 2018 IEEE/ACM 1ST INTERNATIONAL WORKSHOP ON SOFTWARE ENGINEERING FOR AI IN AUTONOMOUS SYSTEMS (SEFAIAS), 2018, : 35 - 38
  • [47] Nav-Q: quantum deep reinforcement learning for collision-free navigation of self-driving cars
    Sinha, Akash
    Macaluso, Antonio
    Klusch, Matthias
    QUANTUM MACHINE INTELLIGENCE, 2025, 7 (01)
  • [48] End-to-End Learning for Lane Keeping of Self-Driving Cars
    Chen, Zhilu
    Huang, Xinming
    2017 28TH IEEE INTELLIGENT VEHICLES SYMPOSIUM (IV 2017), 2017, : 1856 - 1860
  • [49] Predicting Steering Actions for Self-Driving Cars Through Deep Learning
    Ou, Chaojie
    Bedawi, Safaa Mahmoud
    Koesdwiady, Arief B.
    Karray, Fakhri
    2018 IEEE 88TH VEHICULAR TECHNOLOGY CONFERENCE (VTC-FALL), 2018,
  • [50] An Improved Deep Learning Solution for Object Detection in Self-Driving Cars
    Mobahi, Mina
    Sadati, Seyed Hossein
    2020 28TH IRANIAN CONFERENCE ON ELECTRICAL ENGINEERING (ICEE), 2020, : 316 - 320