Double Deep Q-Learning and Faster R-CNN-Based Autonomous Vehicle Navigation and Obstacle Avoidance in Dynamic Environment

被引:19
作者
Bin Issa, Razin [1 ]
Das, Modhumonty [1 ]
Rahman, Md. Saferi [1 ]
Barua, Monika [1 ]
Rhaman, Md. Khalilur [1 ]
Ripon, Kazi Shah Nawaz [2 ]
Alam, Md. Golam Rabiul [1 ]
机构
[1] BRAC Univ, Sch Data & Sci, Dept Comp Sci & Engn, 66 Mohakhali, Dhaka 1212, Bangladesh
[2] Ostfold Univ Coll, Fac Comp Sci, N-1783 Halden, Norway
关键词
autonomous vehicle; reinforcement learning; Double Deep Q Learning; faster R-CNN; object classifier; markov decision process;
D O I
10.3390/s21041468
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
Autonomous vehicle navigation in an unknown dynamic environment is crucial for both supervised- and Reinforcement Learning-based autonomous maneuvering. The cooperative fusion of these two learning approaches has the potential to be an effective mechanism to tackle indefinite environmental dynamics. Most of the state-of-the-art autonomous vehicle navigation systems are trained on a specific mapped model with familiar environmental dynamics. However, this research focuses on the cooperative fusion of supervised and Reinforcement Learning technologies for autonomous navigation of land vehicles in a dynamic and unknown environment. The Faster R-CNN, a supervised learning approach, identifies the ambient environmental obstacles for untroubled maneuver of the autonomous vehicle. Whereas, the training policies of Double Deep Q-Learning, a Reinforcement Learning approach, enable the autonomous agent to learn effective navigation decisions form the dynamic environment. The proposed model is primarily tested in a gaming environment similar to the real-world. It exhibits the overall efficiency and effectiveness in the maneuver of autonomous land vehicles.
引用
收藏
页码:1 / 24
页数:24
相关论文
共 31 条
[1]  
Abadi M, 2016, PROCEEDINGS OF OSDI'16: 12TH USENIX SYMPOSIUM ON OPERATING SYSTEMS DESIGN AND IMPLEMENTATION, P265
[2]  
Abbeel P., 2007, Advances in Neural Information Processing Systems, P2007
[3]   Trajectory-Based Air-Writing Recognition Using Deep Neural Network and Depth Sensor [J].
Alam, Md. Shahinur ;
Kwon, Ki-Chul ;
Alam, Md. Ashraful ;
Abbass, Mohammed Y. ;
Imtiaz, Shariar Md ;
Kim, Nam .
SENSORS, 2020, 20 (02)
[4]  
Bin Issa R, 2020, 2020 34TH INTERNATIONAL CONFERENCE ON INFORMATION NETWORKING (ICOIN 2020), P276, DOI [10.1109/ICOIN48656.2020.9016539, 10.1109/icoin48656.2020.9016539]
[5]  
Carrillo-González José Gerardo, 2015, Ing. invest. y tecnol., V16, P391
[6]  
Chen D, 2020, P AMER CONTR CONF, P4355, DOI [10.23919/ACC45564.2020.9147626, 10.23919/acc45564.2020.9147626]
[7]  
Coggan Melanie, 2004, EXPLORATION EXPLOITA
[8]   Reinforcement learning: The Good, The Bad and The Ugly [J].
Dayana, Peter ;
Niv, Yael .
CURRENT OPINION IN NEUROBIOLOGY, 2008, 18 (02) :185-196
[9]   Scalable Object Detection using Deep Neural Networks [J].
Erhan, Dumitru ;
Szegedy, Christian ;
Toshev, Alexander ;
Anguelov, Dragomir .
2014 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2014, :2155-2162
[10]  
Greenwald A., 2003, CORRELATED Q LEARNIN, V3, P242