Memory-based reinforcement learning algorithm for autonomous exploration in unknown environment

被引:6
作者
Dooraki, Amir Ramezani [1 ]
Lee, Deok Jin [1 ]
机构
[1] Kunsan Natl Univ, Smart Autonomous Syst Lab, Gunsan 54150, South Korea
来源
INTERNATIONAL JOURNAL OF ADVANCED ROBOTIC SYSTEMS | 2018年 / 15卷 / 03期
基金
新加坡国家研究基金会;
关键词
Reinforcement learning; autonomous exploration; adaptive agent; depth map; artificial neural network; sensor fusion; memory-based; obstacle avoidance; ROBOT;
D O I
10.1177/1729881418775849
中图分类号
TP24 [机器人技术];
学科分类号
080202 ; 1405 ;
摘要
In the near future, robots would be seen in almost every area of our life, in different shapes and with different objectives such as entertainment, surveillance, rescue, and navigation. In any shape and with any objective, it is necessary for them to be capable of successful exploration. They should be able to explore efficiently and be able to adapt themselves with changes in their environment. For successful navigation, it is necessary to recognize the difference between similar places of an environment. In order to achieve this goal without increasing the capability of sensors, having a memory is crucial. In this article, an algorithm for autonomous exploration and obstacle avoidance in an unknown environment is proposed. In order to make our self-learner algorithm, a memory-based reinforcement learning method using multilayer neural network is used with the aim of creating an agent having an efficient exploration and obstacle avoidance policy. Furthermore, this agent can automatically adapt itself to the changes of its environment. Finally, in order to test the capability of our algorithm, we have implemented it in a robot similar to a real model, simulated in the robust physics engine simulator of Gazebo.
引用
收藏
页数:10
相关论文
共 19 条
[1]  
[Anonymous], 2017, AUTONOMOUS DEEP LEAR
[2]  
[Anonymous], 2017, POW WORLDS ROB
[3]  
[Anonymous], 1992, THESIS CARNEGIE MELL
[4]  
[Anonymous], 2017, GAZEBO SIMULATOR
[5]  
[Anonymous], 2020, Reinforcement Learning, An Introduction
[6]   Obstacle Avoidance Method for Wheeled Mobile Robots Using Interval Type-2 Fuzzy Neural Network [J].
Kim, Cheol-Joong ;
Chwa, Dongkyoung .
IEEE TRANSACTIONS ON FUZZY SYSTEMS, 2015, 23 (03) :677-687
[7]  
Kominami K, 2012, 2012 IEEE/SICE INTERNATIONAL SYMPOSIUM ON SYSTEM INTEGRATION (SII), P67, DOI 10.1109/SII.2012.6426933
[8]   Quadruped robot obstacle negotiation via reinforcement learning [J].
Lee, Honglak ;
Shen, Yirong ;
Yu, Chih-Han ;
Singh, Gurjeet ;
Ng, Andrew Y. .
2006 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), VOLS 1-10, 2006, :3003-+
[9]   Integrated particle swarm optimization algorithm based obstacle avoidance control design for home service robot [J].
Lin, Chih-Jui ;
Li, Tzuu-Hseng S. ;
Kuo, Ping-Huan ;
Wang, Yin-Hao .
COMPUTERS & ELECTRICAL ENGINEERING, 2016, 56 :748-762
[10]   Human-level control through deep reinforcement learning [J].
Mnih, Volodymyr ;
Kavukcuoglu, Koray ;
Silver, David ;
Rusu, Andrei A. ;
Veness, Joel ;
Bellemare, Marc G. ;
Graves, Alex ;
Riedmiller, Martin ;
Fidjeland, Andreas K. ;
Ostrovski, Georg ;
Petersen, Stig ;
Beattie, Charles ;
Sadik, Amir ;
Antonoglou, Ioannis ;
King, Helen ;
Kumaran, Dharshan ;
Wierstra, Daan ;
Legg, Shane ;
Hassabis, Demis .
NATURE, 2015, 518 (7540) :529-533