Optimal combination of imitation and reinforcement learning for self-driving cars

被引:2
|
作者
Youssef F. [1 ]
Houda B. [1 ]
机构
[1] National School of Computer Science and Systems Analysis (ENSIAS), Mohammed V University, Rabat
来源
Revue d'Intelligence Artificielle | 2019年 / 33卷 / 04期
关键词
Behavioral cloning; Deep reinforcement learning; Expert's trust margin; Prioritized experience replay; Simulation environment; Supervised imitation learning;
D O I
10.18280/ria.330402
中图分类号
学科分类号
摘要
The two steps in human intelligence development, namely, mimicking and tentative application of expertise, are reflected by imitation learning (IL) and reinforcement learning (RL) in artificial intelligence (AI). However, the RL process does not always improve the skills learned from expert demonstrations and enhance the algorithm performance. To solve the problem, this paper puts forward a novel algorithm called optimal combination of imitation and reinforcement learning (OCIRL). First, the concept of deep q-learning from demonstrations (DQfD) was introduced to the actor-critic (A2C) model, creating the A2CfD model. Then, a threshold was estimated from a trained IL model with the same inputs and reward function with the DOfD, and applied to the A2CfD model. The threshold represents the minimum reward that conserves the learned expertise. The resulting A2CfDoC model was trained and tested on self-driving cars in both discrete and continuous environments. The results show that the model outperformed several existing algorithms in terms of speed and accuracy. © 2019 Lavoisier. All rights reserved.
引用
收藏
页码:265 / 273
页数:8
相关论文
共 50 条
  • [1] Deep imitation reinforcement learning for self-driving by vision
    Zou, Qijie
    Xiong, Kang
    Fang, Qiang
    Jiang, Bohan
    CAAI TRANSACTIONS ON INTELLIGENCE TECHNOLOGY, 2021, 6 (04) : 493 - 503
  • [2] Confidence-Aware Reinforcement Learning for Self-Driving Cars
    Cao, Zhong
    Xu, Shaobing
    Peng, Huei
    Yang, Diange
    Zidek, Robert
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2022, 23 (07) : 7419 - 7430
  • [3] A Reinforcement Learning Approach to Smart Lane Changes of Self-driving Cars
    Ye, Fangmin
    Wang, Long
    Wang, Yibing
    Guo, Jingqiu
    Papamichail, Ioannis
    Papageorgiou, Markos
    PROGRESS IN ARTIFICIAL INTELLIGENCE, EPIA 2019, PT I, 2019, 11804 : 559 - 571
  • [4] Self-driving cars
    Becker, Edward
    Becker, Edward, 1600, Society of Tribologists and Lubrication Engineers (77):
  • [5] Self-Driving Cars
    Orbay, Selin
    9TH INTERNATIONAL CONFERENCE THE FUTURE OF EDUCATION, 2019, : 204 - 209
  • [6] SELF-DRIVING CARS
    Kruger, Gary
    SCIENTIFIC AMERICAN, 2016, 315 (04) : 8 - 8
  • [7] Self-Driving Cars
    Daily, Mike
    Medasani, Swarup
    Behringer, Reinhold
    Trivedi, Mohan
    COMPUTER, 2017, 50 (12) : 18 - 23
  • [8] Self-driving Cars
    Andrews, Nicola
    LANDFALL, 2024, (247): : 61 - 61
  • [9] Visualizing the Learning Progress of Self-Driving Cars
    Mund, Sandro
    Frank, Raphael
    Varisteas, Georgios
    State, Radu
    2018 21ST INTERNATIONAL CONFERENCE ON INTELLIGENT TRANSPORTATION SYSTEMS (ITSC), 2018, : 2358 - 2363
  • [10] Hybrid Online POMDP Planning and Deep Reinforcement Learning for Safer Self-Driving Cars
    Pusse, Florian
    Klusch, Matthias
    2019 30TH IEEE INTELLIGENT VEHICLES SYMPOSIUM (IV19), 2019, : 1013 - 1020