Deep Reinforcement Learning For Visual Navigation of Wheeled Mobile Robots

被引:3
|
作者
Nwaonumah, Ezebuugo [1 ]
Samanta, Biswanath [1 ]
机构
[1] Georgia Southern Univ, Dept Mech Engn, Statesboro, GA 30460 USA
来源
IEEE SOUTHEASTCON 2020 | 2020年
关键词
asynchronous advantage actor-critic (A3C); convolutional neural network (CNN); deep neural network (DNN); deep reinforcement learning (DRL); machine learning (ML); mapless navigation; reinforcement learning (RL); ResNet50; robotics; robot operating system (ROS);
D O I
10.1109/southeastcon44009.2020.9249654
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
A study is presented on applying deep reinforcement learning (DRL) for visual navigation of wheeled mobile robots (WMR) in dynamic and unknown environments. Two DRL algorithms, namely, value-learning deep Q-network (DQN) and policy gradient based asynchronous advantage actor critic (A3C), have been considered. RGB (red, green and blue) and depth images have been used as inputs in implementation of both DRL algorithms to generate control commands for autonomous navigation of WMR in simulation environments. The initial DRL networks were generated and trained progressively in OpenAI Gym Gazebo based simulation environments within robot operating system (ROS) framework for a popular target WMR, Kobuki TurtleBot2. A pre-trained deep neural network ResNet50 was used after further training with regrouped objects commonly found in laboratory setting for target-driven mapless visual navigation of Turlebot2 through DRL. The performance of A3C with multiple computation threads (4, 6, and 8) was simulated on a desktop. The navigation performance of DQN and A3C networks, in terms of reward statistics and completion time, was compared in three simulation environments. As expected, A3C with multiple threads (4, 6, and 8) performed better than DQN and the performance of A3C improved with number of threads. Details of the methodology, simulation results are presented and recommendations for future work towards real-time implementation through transfer learning of the DRL models are outlined.
引用
收藏
页数:8
相关论文
共 50 条
  • [21] Deep Reinforcement Learning With Visual Attention for Vehicle Classification
    Zhao, Dongbin
    Chen, Yaran
    Lv, Le
    IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2017, 9 (04) : 356 - 367
  • [22] Visual semantic navigation with real robots
    Gutierrez-alvarez, Carlos
    Rios-Navarro, Pablo
    Flor-Rodriguez-Rabadan, Rafael
    Acevedo-Rodriguez, Francisco Javier
    Lopez-Sastre, Roberto Javier
    APPLIED INTELLIGENCE, 2025, 55 (02)
  • [23] Deep Reinforcement Learning With Multiple Unrelated Rewards for AGV Mapless Navigation
    Cai, Boliang
    Wei, Changyun
    Ji, Ze
    IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2024, : 4323 - 4340
  • [24] Robot Mapless Navigation in VUCA Environments via Deep Reinforcement Learning
    Xue, Bingxin
    Zhou, Fengyu
    Wang, Chaoqun
    Gao, Ming
    Yin, Lei
    IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2025, 72 (01) : 639 - 649
  • [25] Predictive reinforcement learning: map-less navigation method for mobile robot
    Dobriborsci, Dmitrii
    Zashchitin, Roman
    Kakanov, Mikhail
    Aumer, Wolfgang
    Osinenko, Pavel
    JOURNAL OF INTELLIGENT MANUFACTURING, 2023, 35 (8) : 4217 - 4232
  • [26] Mobile Robot Navigation System Using Reinforcement Learning with Path Planning Algorithm
    Andarge, E.W.
    Ordys, A.
    Abebe, Y.M.
    Acta Physica Polonica A, 2024, 146 (04) : 452 - 456
  • [27] Fuzzy situation based navigation of autonomous mobile robot using reinforcement learning
    Guanlao, R
    Musilek, P
    Ahmed, F
    Kaboli, A
    NAFIPS 2004: ANNUAL MEETING OF THE NORTH AMERICAN FUZZY INFORMATION PROCESSING SOCIETY, VOLS 1AND 2: FUZZY SETS IN THE HEART OF THE CANADIAN ROCKIES, 2004, : 820 - 825
  • [28] Control Allocation for Wheeled Mobile Robots Subject to Input Saturation
    Alves, Jhomolos G.
    Lizarralde, Fernando
    Monteiro, Joao C.
    IFAC PAPERSONLINE, 2020, 53 (02): : 3904 - 3909
  • [29] Trajectory Generation for Wheeled Mobile Robots Via Bezier Polynomials
    Sanchez, C. M.
    Sanchez, J. R. G.
    Cervantes, C. Y. S.
    Ortigoza, R. S.
    Guzman, V. M. H.
    Juarez, J. N. A.
    Aranda, M. M.
    IEEE LATIN AMERICA TRANSACTIONS, 2016, 14 (11) : 4482 - 4490
  • [30] GRIMGEP: Learning Progress for Robust Goal Sampling in Visual Deep Reinforcement Learning
    Kovac, Grgur
    Laversanne-Finot, Adrien
    Oudeyer, Pierre-Yves
    IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2023, 15 (03) : 1396 - 1407